var/home/core/zuul-output/0000755000175000017500000000000015115767546014546 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015116004045015465 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005712070315116004035017675 0ustar rootrootDec 09 09:47:22 crc systemd[1]: Starting Kubernetes Kubelet... Dec 09 09:47:22 crc restorecon[4694]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:22 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 09:47:23 crc restorecon[4694]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 09:47:23 crc restorecon[4694]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 09 09:47:23 crc kubenswrapper[4824]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 09 09:47:23 crc kubenswrapper[4824]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 09 09:47:23 crc kubenswrapper[4824]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 09 09:47:23 crc kubenswrapper[4824]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 09 09:47:23 crc kubenswrapper[4824]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 09 09:47:23 crc kubenswrapper[4824]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.777703 4824 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780695 4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780709 4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780720 4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780724 4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780728 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780732 4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780736 4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780741 4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780745 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780750 4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780754 4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780757 4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780761 4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780764 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780768 4824 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780771 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780775 4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780791 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780794 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780798 4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780802 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780806 4824 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780809 4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780813 4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780816 4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780820 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780824 4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780828 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780832 4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780835 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780839 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780842 4824 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780846 4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780850 4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780853 4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780858 4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780862 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780868 4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780873 4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780877 4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780881 4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780885 4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780888 4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780892 4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780896 4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780899 4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780903 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780907 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780911 4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780915 4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780918 4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780922 4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780926 4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780929 4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780933 4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780936 4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780941 4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780945 4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780948 4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780952 4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780955 4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780959 4824 feature_gate.go:330] unrecognized feature gate: Example Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780962 4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780966 4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780969 4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780973 4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780978 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780983 4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780988 4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780993 4824 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.780997 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781252 4824 flags.go:64] FLAG: --address="0.0.0.0" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781264 4824 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781271 4824 flags.go:64] FLAG: --anonymous-auth="true" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781278 4824 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781283 4824 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781287 4824 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781293 4824 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781298 4824 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781303 4824 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781307 4824 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781312 4824 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781316 4824 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781320 4824 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781324 4824 flags.go:64] FLAG: --cgroup-root="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781328 4824 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781332 4824 flags.go:64] FLAG: --client-ca-file="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781336 4824 flags.go:64] FLAG: --cloud-config="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781340 4824 flags.go:64] FLAG: --cloud-provider="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781344 4824 flags.go:64] FLAG: --cluster-dns="[]" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781350 4824 flags.go:64] FLAG: --cluster-domain="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781354 4824 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781358 4824 flags.go:64] FLAG: --config-dir="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781362 4824 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781367 4824 flags.go:64] FLAG: --container-log-max-files="5" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781372 4824 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781376 4824 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781381 4824 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781385 4824 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781390 4824 flags.go:64] FLAG: --contention-profiling="false" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781394 4824 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781400 4824 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781404 4824 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781409 4824 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781414 4824 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781419 4824 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781423 4824 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781427 4824 flags.go:64] FLAG: --enable-load-reader="false" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781431 4824 flags.go:64] FLAG: --enable-server="true" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781435 4824 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781441 4824 flags.go:64] FLAG: --event-burst="100" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781446 4824 flags.go:64] FLAG: --event-qps="50" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781450 4824 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781455 4824 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781459 4824 flags.go:64] FLAG: --eviction-hard="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781465 4824 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781469 4824 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781473 4824 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781477 4824 flags.go:64] FLAG: --eviction-soft="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781481 4824 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781485 4824 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781489 4824 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781493 4824 flags.go:64] FLAG: --experimental-mounter-path="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781498 4824 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781502 4824 flags.go:64] FLAG: --fail-swap-on="true" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781506 4824 flags.go:64] FLAG: --feature-gates="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781511 4824 flags.go:64] FLAG: --file-check-frequency="20s" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781515 4824 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781519 4824 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781524 4824 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781528 4824 flags.go:64] FLAG: --healthz-port="10248" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781533 4824 flags.go:64] FLAG: --help="false" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781538 4824 flags.go:64] FLAG: --hostname-override="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781542 4824 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781546 4824 flags.go:64] FLAG: --http-check-frequency="20s" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781551 4824 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781555 4824 flags.go:64] FLAG: --image-credential-provider-config="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781558 4824 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781562 4824 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781567 4824 flags.go:64] FLAG: --image-service-endpoint="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781571 4824 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781575 4824 flags.go:64] FLAG: --kube-api-burst="100" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781579 4824 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781583 4824 flags.go:64] FLAG: --kube-api-qps="50" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781587 4824 flags.go:64] FLAG: --kube-reserved="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781591 4824 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781596 4824 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781601 4824 flags.go:64] FLAG: --kubelet-cgroups="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781605 4824 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781609 4824 flags.go:64] FLAG: --lock-file="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781613 4824 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781617 4824 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781621 4824 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781627 4824 flags.go:64] FLAG: --log-json-split-stream="false" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781631 4824 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781635 4824 flags.go:64] FLAG: --log-text-split-stream="false" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781640 4824 flags.go:64] FLAG: --logging-format="text" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781644 4824 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781648 4824 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781652 4824 flags.go:64] FLAG: --manifest-url="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781656 4824 flags.go:64] FLAG: --manifest-url-header="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781662 4824 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781666 4824 flags.go:64] FLAG: --max-open-files="1000000" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781672 4824 flags.go:64] FLAG: --max-pods="110" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781676 4824 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781680 4824 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781685 4824 flags.go:64] FLAG: --memory-manager-policy="None" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781689 4824 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781693 4824 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781697 4824 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781701 4824 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781710 4824 flags.go:64] FLAG: --node-status-max-images="50" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781714 4824 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781718 4824 flags.go:64] FLAG: --oom-score-adj="-999" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781722 4824 flags.go:64] FLAG: --pod-cidr="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781726 4824 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781732 4824 flags.go:64] FLAG: --pod-manifest-path="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781736 4824 flags.go:64] FLAG: --pod-max-pids="-1" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781741 4824 flags.go:64] FLAG: --pods-per-core="0" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781745 4824 flags.go:64] FLAG: --port="10250" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781749 4824 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781754 4824 flags.go:64] FLAG: --provider-id="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781763 4824 flags.go:64] FLAG: --qos-reserved="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781767 4824 flags.go:64] FLAG: --read-only-port="10255" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781771 4824 flags.go:64] FLAG: --register-node="true" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781789 4824 flags.go:64] FLAG: --register-schedulable="true" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781794 4824 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781801 4824 flags.go:64] FLAG: --registry-burst="10" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781806 4824 flags.go:64] FLAG: --registry-qps="5" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781810 4824 flags.go:64] FLAG: --reserved-cpus="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781814 4824 flags.go:64] FLAG: --reserved-memory="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781819 4824 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781824 4824 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781828 4824 flags.go:64] FLAG: --rotate-certificates="false" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781832 4824 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781836 4824 flags.go:64] FLAG: --runonce="false" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781840 4824 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781845 4824 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781849 4824 flags.go:64] FLAG: --seccomp-default="false" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781853 4824 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781857 4824 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781862 4824 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781866 4824 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781870 4824 flags.go:64] FLAG: --storage-driver-password="root" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781874 4824 flags.go:64] FLAG: --storage-driver-secure="false" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781878 4824 flags.go:64] FLAG: --storage-driver-table="stats" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781883 4824 flags.go:64] FLAG: --storage-driver-user="root" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781887 4824 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781892 4824 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781896 4824 flags.go:64] FLAG: --system-cgroups="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781900 4824 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781906 4824 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781911 4824 flags.go:64] FLAG: --tls-cert-file="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781915 4824 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781921 4824 flags.go:64] FLAG: --tls-min-version="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781925 4824 flags.go:64] FLAG: --tls-private-key-file="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781930 4824 flags.go:64] FLAG: --topology-manager-policy="none" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781934 4824 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781939 4824 flags.go:64] FLAG: --topology-manager-scope="container" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781944 4824 flags.go:64] FLAG: --v="2" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781950 4824 flags.go:64] FLAG: --version="false" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781955 4824 flags.go:64] FLAG: --vmodule="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781961 4824 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.781965 4824 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782104 4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782109 4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782114 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782119 4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782123 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782130 4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782134 4824 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782138 4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782143 4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782148 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782151 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782156 4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782161 4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782165 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782169 4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782173 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782177 4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782180 4824 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782184 4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782188 4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782193 4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782198 4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782202 4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782207 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782212 4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782217 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782221 4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782226 4824 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782230 4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782234 4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782240 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782244 4824 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782249 4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782253 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782257 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782261 4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782266 4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782272 4824 feature_gate.go:330] unrecognized feature gate: Example Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782276 4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782280 4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782284 4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782289 4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782293 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782297 4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782301 4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782305 4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782310 4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782314 4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782318 4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782322 4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782326 4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782331 4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782335 4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782339 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782343 4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782347 4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782352 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782357 4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782361 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782365 4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782369 4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782374 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782378 4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782382 4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782387 4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782391 4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782397 4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782402 4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782406 4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782411 4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.782415 4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.782580 4824 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.789423 4824 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.789460 4824 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789528 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789536 4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789542 4824 feature_gate.go:330] unrecognized feature gate: Example Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789547 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789551 4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789555 4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789559 4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789564 4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789568 4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789572 4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789576 4824 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789580 4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789585 4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789589 4824 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789593 4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789597 4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789602 4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789608 4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789612 4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789616 4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789620 4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789625 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789629 4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789633 4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789638 4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789643 4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789647 4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789652 4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789656 4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789660 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789664 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789668 4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789672 4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789677 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789681 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789685 4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789689 4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789694 4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789698 4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789702 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789707 4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789711 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789715 4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789720 4824 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789723 4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789728 4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789732 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789737 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789741 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789744 4824 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789748 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789752 4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789755 4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789759 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789762 4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789766 4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789769 4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789774 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789790 4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789794 4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789798 4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789801 4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789806 4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789811 4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789815 4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789818 4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789822 4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789825 4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789829 4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789833 4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789837 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.789844 4824 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789959 4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789968 4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789972 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789977 4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789981 4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789985 4824 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789989 4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789992 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.789996 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790036 4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790041 4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790044 4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790048 4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790052 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790056 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790060 4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790063 4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790069 4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790075 4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790081 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790086 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790092 4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790098 4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790107 4824 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790114 4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790120 4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790125 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790131 4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790136 4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790141 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790145 4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790150 4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790154 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790158 4824 feature_gate.go:330] unrecognized feature gate: Example Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790162 4824 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790166 4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790170 4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790174 4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790177 4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790181 4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790185 4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790189 4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790193 4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790197 4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790201 4824 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790205 4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790209 4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790213 4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790218 4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790222 4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790226 4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790229 4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790234 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790238 4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790243 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790246 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790250 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790256 4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790259 4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790263 4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790267 4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790272 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790277 4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790281 4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790284 4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790288 4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790292 4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790297 4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790301 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790304 4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.790308 4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.790315 4824 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.790683 4824 server.go:940] "Client rotation is on, will bootstrap in background" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.793976 4824 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.794138 4824 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.794814 4824 server.go:997] "Starting client certificate rotation" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.794840 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.795018 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-01 01:44:07.39716807 +0000 UTC Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.795103 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.803302 4824 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.805021 4824 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 09 09:47:23 crc kubenswrapper[4824]: E1209 09:47:23.805232 4824 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.128:6443: connect: connection refused" logger="UnhandledError" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.812442 4824 log.go:25] "Validated CRI v1 runtime API" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.826319 4824 log.go:25] "Validated CRI v1 image API" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.827568 4824 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.830926 4824 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-09-09-42-48-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.830952 4824 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.845039 4824 manager.go:217] Machine: {Timestamp:2025-12-09 09:47:23.843841824 +0000 UTC m=+0.178346511 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09 BootID:047c928d-3ca9-4a45-942a-c245de14f559 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:51:5b:4b Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:51:5b:4b Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:53:24:c0 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:86:55:79 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:81:28:18 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:80:3e:57 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ba:53:7b:fa:99:71 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:ba:bc:93:75:4c:6f Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.845322 4824 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.845437 4824 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.846212 4824 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.846356 4824 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.846384 4824 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.846567 4824 topology_manager.go:138] "Creating topology manager with none policy" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.846575 4824 container_manager_linux.go:303] "Creating device plugin manager" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.846746 4824 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.846792 4824 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.846943 4824 state_mem.go:36] "Initialized new in-memory state store" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.847017 4824 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.847647 4824 kubelet.go:418] "Attempting to sync node with API server" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.847665 4824 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.847689 4824 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.847705 4824 kubelet.go:324] "Adding apiserver pod source" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.847718 4824 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.852023 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.128:6443: connect: connection refused Dec 09 09:47:23 crc kubenswrapper[4824]: E1209 09:47:23.852093 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.128:6443: connect: connection refused" logger="UnhandledError" Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.852571 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.128:6443: connect: connection refused Dec 09 09:47:23 crc kubenswrapper[4824]: E1209 09:47:23.852611 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.128:6443: connect: connection refused" logger="UnhandledError" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.853196 4824 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.853896 4824 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.855434 4824 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.856365 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.856388 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.856396 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.856402 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.856415 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.856422 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.856430 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.856443 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.856452 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.856462 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.856488 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.856498 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.856908 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.857396 4824 server.go:1280] "Started kubelet" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.857905 4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.128:6443: connect: connection refused Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.858210 4824 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.858208 4824 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.859490 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.859527 4824 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.859595 4824 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 09 09:47:23 crc systemd[1]: Started Kubernetes Kubelet. Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.859751 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 08:30:15.60813384 +0000 UTC Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.859823 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 838h42m51.748314428s for next certificate rotation Dec 09 09:47:23 crc kubenswrapper[4824]: E1209 09:47:23.859893 4824 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.860247 4824 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.860267 4824 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.860321 4824 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.860828 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.128:6443: connect: connection refused Dec 09 09:47:23 crc kubenswrapper[4824]: E1209 09:47:23.860863 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.128:6443: connect: connection refused" interval="200ms" Dec 09 09:47:23 crc kubenswrapper[4824]: E1209 09:47:23.860887 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.128:6443: connect: connection refused" logger="UnhandledError" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.861194 4824 factory.go:55] Registering systemd factory Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.861220 4824 factory.go:221] Registration of the systemd container factory successfully Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.861703 4824 factory.go:153] Registering CRI-O factory Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.861732 4824 factory.go:221] Registration of the crio container factory successfully Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.861821 4824 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.861870 4824 factory.go:103] Registering Raw factory Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.861888 4824 manager.go:1196] Started watching for new ooms in manager Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.862523 4824 manager.go:319] Starting recovery of all containers Dec 09 09:47:23 crc kubenswrapper[4824]: E1209 09:47:23.861489 4824 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.128:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187f8306b97a1d9d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-09 09:47:23.857354141 +0000 UTC m=+0.191858818,LastTimestamp:2025-12-09 09:47:23.857354141 +0000 UTC m=+0.191858818,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.863098 4824 server.go:460] "Adding debug handlers to kubelet server" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.873639 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.873754 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.873770 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.873808 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.873820 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.873831 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.873841 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.873853 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.873870 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.873884 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.873895 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.873952 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.873975 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.873992 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874039 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874059 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874074 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874087 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874185 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874203 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874218 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874232 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874247 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874263 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874300 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874320 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874340 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874357 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874377 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874393 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874408 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874429 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874443 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874460 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874483 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874500 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874516 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874532 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874548 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874565 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874579 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874594 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874611 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874629 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874646 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874665 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874680 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874695 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874708 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874719 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874728 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874740 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874753 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874765 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874797 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874813 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874826 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874839 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874851 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874867 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874882 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874897 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874914 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874928 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874942 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874958 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874970 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874983 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.874995 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875009 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875019 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875033 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875046 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875058 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875069 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875080 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875093 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875107 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875119 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875129 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875139 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875150 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875161 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875172 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875183 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875194 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875210 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875222 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875233 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875244 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875255 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875267 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875281 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875292 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875302 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875313 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875323 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875334 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875346 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875357 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875368 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875410 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875423 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875438 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875453 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875465 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875476 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875489 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875502 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875518 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875529 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875539 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875550 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875559 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875570 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875582 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875591 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875602 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875613 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875625 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875637 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875651 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875663 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875674 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875686 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875697 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875707 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875719 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875730 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875742 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875752 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875762 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875772 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875804 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875816 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875826 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875838 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875848 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875860 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.875869 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877019 4824 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877104 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877134 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877156 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877177 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877199 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877216 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877233 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877256 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877271 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877288 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877308 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877361 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877391 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877410 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877427 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877446 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877459 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877474 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877488 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877502 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877518 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877530 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877542 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877554 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877566 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877579 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877590 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877604 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877616 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877627 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877642 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877657 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877668 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877680 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877693 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877706 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877717 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877738 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877752 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877764 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877791 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877808 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877821 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877833 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877846 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877858 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877873 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877886 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877898 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877937 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877951 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877964 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877976 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.877989 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.878002 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.878016 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.878029 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.878040 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.878053 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.878068 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.878081 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.878093 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.878106 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.878121 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.878134 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.878147 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.878164 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.878175 4824 reconstruct.go:97] "Volume reconstruction finished" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.878184 4824 reconciler.go:26] "Reconciler: start to sync state" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.885128 4824 manager.go:324] Recovery completed Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.896683 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.899914 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.899979 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.899996 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.902074 4824 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.902910 4824 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.902959 4824 state_mem.go:36] "Initialized new in-memory state store" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.907800 4824 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.909081 4824 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.909136 4824 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 09 09:47:23 crc kubenswrapper[4824]: I1209 09:47:23.909180 4824 kubelet.go:2335] "Starting kubelet main sync loop" Dec 09 09:47:23 crc kubenswrapper[4824]: E1209 09:47:23.909259 4824 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 09 09:47:23 crc kubenswrapper[4824]: W1209 09:47:23.909828 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.128:6443: connect: connection refused Dec 09 09:47:23 crc kubenswrapper[4824]: E1209 09:47:23.909911 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.128:6443: connect: connection refused" logger="UnhandledError" Dec 09 09:47:23 crc kubenswrapper[4824]: E1209 09:47:23.961017 4824 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 09 09:47:24 crc kubenswrapper[4824]: E1209 09:47:24.009828 4824 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.010118 4824 policy_none.go:49] "None policy: Start" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.012322 4824 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.012358 4824 state_mem.go:35] "Initializing new in-memory state store" Dec 09 09:47:24 crc kubenswrapper[4824]: E1209 09:47:24.061146 4824 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 09 09:47:24 crc kubenswrapper[4824]: E1209 09:47:24.061622 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.128:6443: connect: connection refused" interval="400ms" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.067692 4824 manager.go:334] "Starting Device Plugin manager" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.067986 4824 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.068014 4824 server.go:79] "Starting device plugin registration server" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.068447 4824 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.068470 4824 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.068867 4824 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.070194 4824 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.070229 4824 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 09 09:47:24 crc kubenswrapper[4824]: E1209 09:47:24.078977 4824 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.169575 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.171271 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.171323 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.171335 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.171364 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 09:47:24 crc kubenswrapper[4824]: E1209 09:47:24.171965 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.128:6443: connect: connection refused" node="crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.210947 4824 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.211087 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.212132 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.212160 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.212170 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.212295 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.212563 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.212651 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.213139 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.213188 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.213199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.213369 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.213629 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.213695 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.214035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.214069 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.214082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.214362 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.214383 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.214401 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.214411 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.214387 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.214479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.214522 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.214612 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.214644 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.215133 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.215152 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.215162 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.215206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.215223 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.215235 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.215258 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.215439 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.215475 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.216012 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.216034 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.216046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.216044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.216179 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.216191 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.216357 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.216382 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.216952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.216980 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.216992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.281963 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.282055 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.282088 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.282112 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.282199 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.282226 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.282248 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.282274 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.282314 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.282348 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.282386 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.282486 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.282536 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.282558 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.282577 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: W1209 09:47:24.284135 4824 helpers.go:245] readString: Failed to read "/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/cpuset.cpus.effective": read /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/cpuset.cpus.effective: no such device Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.372861 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.374569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.374639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.374656 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.374683 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 09:47:24 crc kubenswrapper[4824]: E1209 09:47:24.375268 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.128:6443: connect: connection refused" node="crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.383572 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.383637 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.383666 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.383689 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.383710 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.383728 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.383750 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.383773 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.383820 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.383839 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.383858 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.383879 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.383863 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.383930 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.384002 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.384038 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.383900 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.384077 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.384109 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.384161 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.383955 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.384027 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.384074 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.384084 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.383980 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.384114 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.384027 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.384271 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.384012 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.384388 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: E1209 09:47:24.462576 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.128:6443: connect: connection refused" interval="800ms" Dec 09 09:47:24 crc kubenswrapper[4824]: E1209 09:47:24.535340 4824 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.128:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187f8306b97a1d9d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-09 09:47:23.857354141 +0000 UTC m=+0.191858818,LastTimestamp:2025-12-09 09:47:23.857354141 +0000 UTC m=+0.191858818,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.552740 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.561402 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.580640 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: W1209 09:47:24.582492 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-03a3943892ffc1662003dc86b636aac9637ff624eac699484f62f8c32fb5288b WatchSource:0}: Error finding container 03a3943892ffc1662003dc86b636aac9637ff624eac699484f62f8c32fb5288b: Status 404 returned error can't find the container with id 03a3943892ffc1662003dc86b636aac9637ff624eac699484f62f8c32fb5288b Dec 09 09:47:24 crc kubenswrapper[4824]: W1209 09:47:24.585213 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-4660f10779533927279cd6136add740e74789780a51eeb041088e37575ef623c WatchSource:0}: Error finding container 4660f10779533927279cd6136add740e74789780a51eeb041088e37575ef623c: Status 404 returned error can't find the container with id 4660f10779533927279cd6136add740e74789780a51eeb041088e37575ef623c Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.598563 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.603721 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 09:47:24 crc kubenswrapper[4824]: W1209 09:47:24.652872 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-9507f9be04d619993a369e414844cb63954d51eeaed2de95e15308237c6cb6a1 WatchSource:0}: Error finding container 9507f9be04d619993a369e414844cb63954d51eeaed2de95e15308237c6cb6a1: Status 404 returned error can't find the container with id 9507f9be04d619993a369e414844cb63954d51eeaed2de95e15308237c6cb6a1 Dec 09 09:47:24 crc kubenswrapper[4824]: W1209 09:47:24.654625 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-c25198e912e5392027e1758c7fd6767c748c33800d1b9f2dab7b1de8b0f0b916 WatchSource:0}: Error finding container c25198e912e5392027e1758c7fd6767c748c33800d1b9f2dab7b1de8b0f0b916: Status 404 returned error can't find the container with id c25198e912e5392027e1758c7fd6767c748c33800d1b9f2dab7b1de8b0f0b916 Dec 09 09:47:24 crc kubenswrapper[4824]: W1209 09:47:24.726336 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.128:6443: connect: connection refused Dec 09 09:47:24 crc kubenswrapper[4824]: E1209 09:47:24.726446 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.128:6443: connect: connection refused" logger="UnhandledError" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.776105 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.778276 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.778344 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.778369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.778408 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 09:47:24 crc kubenswrapper[4824]: E1209 09:47:24.779149 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.128:6443: connect: connection refused" node="crc" Dec 09 09:47:24 crc kubenswrapper[4824]: W1209 09:47:24.785167 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.128:6443: connect: connection refused Dec 09 09:47:24 crc kubenswrapper[4824]: E1209 09:47:24.785295 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.128:6443: connect: connection refused" logger="UnhandledError" Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.858627 4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.128:6443: connect: connection refused Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.914074 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9507f9be04d619993a369e414844cb63954d51eeaed2de95e15308237c6cb6a1"} Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.915189 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"23045077c158dd995f3039bfb8c7c44cd03e9a04a10e38df701cdb14c1225376"} Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.916832 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"03a3943892ffc1662003dc86b636aac9637ff624eac699484f62f8c32fb5288b"} Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.918368 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"4660f10779533927279cd6136add740e74789780a51eeb041088e37575ef623c"} Dec 09 09:47:24 crc kubenswrapper[4824]: I1209 09:47:24.919951 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c25198e912e5392027e1758c7fd6767c748c33800d1b9f2dab7b1de8b0f0b916"} Dec 09 09:47:25 crc kubenswrapper[4824]: W1209 09:47:25.066541 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.128:6443: connect: connection refused Dec 09 09:47:25 crc kubenswrapper[4824]: E1209 09:47:25.066648 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.128:6443: connect: connection refused" logger="UnhandledError" Dec 09 09:47:25 crc kubenswrapper[4824]: W1209 09:47:25.169439 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.128:6443: connect: connection refused Dec 09 09:47:25 crc kubenswrapper[4824]: E1209 09:47:25.169524 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.128:6443: connect: connection refused" logger="UnhandledError" Dec 09 09:47:25 crc kubenswrapper[4824]: E1209 09:47:25.263734 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.128:6443: connect: connection refused" interval="1.6s" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.579681 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.592936 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.592974 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.592983 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.593012 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 09:47:25 crc kubenswrapper[4824]: E1209 09:47:25.593506 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.128:6443: connect: connection refused" node="crc" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.859304 4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.128:6443: connect: connection refused Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.926515 4824 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="2d2f673541c034e0520cbed8045424520dddddf04d26bc5f6f2c9c9accde59e8" exitCode=0 Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.926735 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.926741 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"2d2f673541c034e0520cbed8045424520dddddf04d26bc5f6f2c9c9accde59e8"} Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.928033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.928065 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.928075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.929959 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26"} Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.930179 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68"} Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.930208 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0"} Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.935098 4824 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="8eb98ce1943d5692f7e6d51befb0c05115ee403be72134f40f651a80bf0a378b" exitCode=0 Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.935174 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"8eb98ce1943d5692f7e6d51befb0c05115ee403be72134f40f651a80bf0a378b"} Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.935250 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.936501 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.936536 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.936549 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.937474 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b" exitCode=0 Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.937584 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b"} Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.937628 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.938719 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.938752 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.938761 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.941061 4824 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c" exitCode=0 Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.941286 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c"} Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.941616 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.943347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.943457 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.943475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.946618 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.951936 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.951999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.952016 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:25 crc kubenswrapper[4824]: I1209 09:47:25.957648 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 09 09:47:25 crc kubenswrapper[4824]: E1209 09:47:25.958816 4824 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.128:6443: connect: connection refused" logger="UnhandledError" Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.946288 4824 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52" exitCode=0 Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.946366 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52"} Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.946468 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.947827 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.947869 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.947880 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.948953 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.949220 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"1deff21cac9a0d3e7e4def13068ef26370bedd2aa343853fb6d19a0a1179d543"} Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.949571 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.949602 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.949614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.959568 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154"} Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.959588 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.960409 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.960439 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.960450 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.962506 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.962962 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"17601a452dc4369a533806384d58f733e396e255f5741b38004cae98a44cfe55"} Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.963029 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2add2743e390cf2b91c9b020de38fc6df392fee9e5aa1d59bd1ca9471d9d8d79"} Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.963045 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e1e6cfeb997cafb7ec7f6d829e1ff2e2d95e4b343c422194a5290d222ea53a7a"} Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.963414 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.963451 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.963489 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.966285 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90"} Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.966327 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c"} Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.966345 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8"} Dec 09 09:47:26 crc kubenswrapper[4824]: I1209 09:47:26.966359 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39"} Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.194007 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.195595 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.195630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.195639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.195667 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.975849 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042"} Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.975886 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.977311 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.977349 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.977369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.978040 4824 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2" exitCode=0 Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.978129 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2"} Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.978167 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.978280 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.978339 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.978369 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.978340 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.979104 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.979127 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.979136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.979908 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.979936 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.979954 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.979962 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.979939 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.980037 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.980052 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.980075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:27 crc kubenswrapper[4824]: I1209 09:47:27.980086 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:28 crc kubenswrapper[4824]: I1209 09:47:28.984232 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53"} Dec 09 09:47:28 crc kubenswrapper[4824]: I1209 09:47:28.984322 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 09:47:28 crc kubenswrapper[4824]: I1209 09:47:28.984942 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:28 crc kubenswrapper[4824]: I1209 09:47:28.984363 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:28 crc kubenswrapper[4824]: I1209 09:47:28.986351 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:28 crc kubenswrapper[4824]: I1209 09:47:28.986366 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:28 crc kubenswrapper[4824]: I1209 09:47:28.986397 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:28 crc kubenswrapper[4824]: I1209 09:47:28.986402 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:28 crc kubenswrapper[4824]: I1209 09:47:28.986410 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:28 crc kubenswrapper[4824]: I1209 09:47:28.986418 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:29 crc kubenswrapper[4824]: I1209 09:47:29.608979 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:47:29 crc kubenswrapper[4824]: I1209 09:47:29.933427 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 09:47:29 crc kubenswrapper[4824]: I1209 09:47:29.933596 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:29 crc kubenswrapper[4824]: I1209 09:47:29.934836 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:29 crc kubenswrapper[4824]: I1209 09:47:29.934897 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:29 crc kubenswrapper[4824]: I1209 09:47:29.934910 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:29 crc kubenswrapper[4824]: I1209 09:47:29.990422 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c"} Dec 09 09:47:29 crc kubenswrapper[4824]: I1209 09:47:29.990497 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:29 crc kubenswrapper[4824]: I1209 09:47:29.990495 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252"} Dec 09 09:47:29 crc kubenswrapper[4824]: I1209 09:47:29.990597 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650"} Dec 09 09:47:29 crc kubenswrapper[4824]: I1209 09:47:29.990626 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a"} Dec 09 09:47:29 crc kubenswrapper[4824]: I1209 09:47:29.990453 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 09:47:29 crc kubenswrapper[4824]: I1209 09:47:29.990671 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:29 crc kubenswrapper[4824]: I1209 09:47:29.991293 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:29 crc kubenswrapper[4824]: I1209 09:47:29.991333 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:29 crc kubenswrapper[4824]: I1209 09:47:29.991349 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:29 crc kubenswrapper[4824]: I1209 09:47:29.991486 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:29 crc kubenswrapper[4824]: I1209 09:47:29.991511 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:29 crc kubenswrapper[4824]: I1209 09:47:29.991524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:30 crc kubenswrapper[4824]: I1209 09:47:30.238584 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 09 09:47:30 crc kubenswrapper[4824]: I1209 09:47:30.992578 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:30 crc kubenswrapper[4824]: I1209 09:47:30.994129 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:30 crc kubenswrapper[4824]: I1209 09:47:30.994167 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:30 crc kubenswrapper[4824]: I1209 09:47:30.994176 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:31 crc kubenswrapper[4824]: I1209 09:47:31.002508 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:47:31 crc kubenswrapper[4824]: I1209 09:47:31.002713 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:31 crc kubenswrapper[4824]: I1209 09:47:31.003770 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:31 crc kubenswrapper[4824]: I1209 09:47:31.003828 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:31 crc kubenswrapper[4824]: I1209 09:47:31.003842 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:31 crc kubenswrapper[4824]: I1209 09:47:31.370323 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 09:47:31 crc kubenswrapper[4824]: I1209 09:47:31.370516 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:31 crc kubenswrapper[4824]: I1209 09:47:31.372168 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:31 crc kubenswrapper[4824]: I1209 09:47:31.372230 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:31 crc kubenswrapper[4824]: I1209 09:47:31.372246 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:32 crc kubenswrapper[4824]: I1209 09:47:32.775379 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:47:32 crc kubenswrapper[4824]: I1209 09:47:32.775614 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:32 crc kubenswrapper[4824]: I1209 09:47:32.776871 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:32 crc kubenswrapper[4824]: I1209 09:47:32.776908 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:32 crc kubenswrapper[4824]: I1209 09:47:32.776920 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:33 crc kubenswrapper[4824]: I1209 09:47:33.534029 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 09 09:47:33 crc kubenswrapper[4824]: I1209 09:47:33.534300 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:33 crc kubenswrapper[4824]: I1209 09:47:33.535550 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:33 crc kubenswrapper[4824]: I1209 09:47:33.535610 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:33 crc kubenswrapper[4824]: I1209 09:47:33.535629 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:34 crc kubenswrapper[4824]: E1209 09:47:34.079190 4824 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 09 09:47:34 crc kubenswrapper[4824]: I1209 09:47:34.337487 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 09:47:34 crc kubenswrapper[4824]: I1209 09:47:34.337696 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:34 crc kubenswrapper[4824]: I1209 09:47:34.339443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:34 crc kubenswrapper[4824]: I1209 09:47:34.339487 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:34 crc kubenswrapper[4824]: I1209 09:47:34.339500 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:34 crc kubenswrapper[4824]: I1209 09:47:34.370908 4824 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 09:47:34 crc kubenswrapper[4824]: I1209 09:47:34.371044 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 09:47:34 crc kubenswrapper[4824]: I1209 09:47:34.936931 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 09:47:34 crc kubenswrapper[4824]: I1209 09:47:34.942299 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 09:47:35 crc kubenswrapper[4824]: I1209 09:47:35.002250 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:35 crc kubenswrapper[4824]: I1209 09:47:35.003774 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:35 crc kubenswrapper[4824]: I1209 09:47:35.003843 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:35 crc kubenswrapper[4824]: I1209 09:47:35.003859 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:35 crc kubenswrapper[4824]: I1209 09:47:35.006544 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 09:47:35 crc kubenswrapper[4824]: I1209 09:47:35.824422 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 09 09:47:35 crc kubenswrapper[4824]: I1209 09:47:35.824650 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:35 crc kubenswrapper[4824]: I1209 09:47:35.826315 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:35 crc kubenswrapper[4824]: I1209 09:47:35.826369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:35 crc kubenswrapper[4824]: I1209 09:47:35.826383 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:36 crc kubenswrapper[4824]: I1209 09:47:36.004087 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:36 crc kubenswrapper[4824]: I1209 09:47:36.005151 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:36 crc kubenswrapper[4824]: I1209 09:47:36.005188 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:36 crc kubenswrapper[4824]: I1209 09:47:36.005201 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:36 crc kubenswrapper[4824]: I1209 09:47:36.860166 4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 09 09:47:36 crc kubenswrapper[4824]: E1209 09:47:36.865350 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 09 09:47:36 crc kubenswrapper[4824]: W1209 09:47:36.877242 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 09 09:47:36 crc kubenswrapper[4824]: I1209 09:47:36.877351 4824 trace.go:236] Trace[1982433757]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Dec-2025 09:47:26.876) (total time: 10001ms): Dec 09 09:47:36 crc kubenswrapper[4824]: Trace[1982433757]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10000ms (09:47:36.877) Dec 09 09:47:36 crc kubenswrapper[4824]: Trace[1982433757]: [10.001144096s] [10.001144096s] END Dec 09 09:47:36 crc kubenswrapper[4824]: E1209 09:47:36.877382 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 09 09:47:36 crc kubenswrapper[4824]: W1209 09:47:36.933578 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 09 09:47:36 crc kubenswrapper[4824]: I1209 09:47:36.933700 4824 trace.go:236] Trace[1105031672]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Dec-2025 09:47:26.932) (total time: 10001ms): Dec 09 09:47:36 crc kubenswrapper[4824]: Trace[1105031672]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (09:47:36.933) Dec 09 09:47:36 crc kubenswrapper[4824]: Trace[1105031672]: [10.001398831s] [10.001398831s] END Dec 09 09:47:36 crc kubenswrapper[4824]: E1209 09:47:36.933726 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 09 09:47:37 crc kubenswrapper[4824]: I1209 09:47:37.006781 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:37 crc kubenswrapper[4824]: I1209 09:47:37.007723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:37 crc kubenswrapper[4824]: I1209 09:47:37.007750 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:37 crc kubenswrapper[4824]: I1209 09:47:37.007760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:37 crc kubenswrapper[4824]: E1209 09:47:37.196365 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 09 09:47:37 crc kubenswrapper[4824]: W1209 09:47:37.226594 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 09 09:47:37 crc kubenswrapper[4824]: I1209 09:47:37.226698 4824 trace.go:236] Trace[1394742773]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Dec-2025 09:47:27.225) (total time: 10001ms): Dec 09 09:47:37 crc kubenswrapper[4824]: Trace[1394742773]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (09:47:37.226) Dec 09 09:47:37 crc kubenswrapper[4824]: Trace[1394742773]: [10.001222287s] [10.001222287s] END Dec 09 09:47:37 crc kubenswrapper[4824]: E1209 09:47:37.226723 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 09 09:47:37 crc kubenswrapper[4824]: I1209 09:47:37.842373 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 09 09:47:37 crc kubenswrapper[4824]: I1209 09:47:37.842446 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 09 09:47:37 crc kubenswrapper[4824]: I1209 09:47:37.846427 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 09 09:47:37 crc kubenswrapper[4824]: I1209 09:47:37.846486 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 09 09:47:39 crc kubenswrapper[4824]: I1209 09:47:39.615513 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:47:39 crc kubenswrapper[4824]: I1209 09:47:39.615754 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:39 crc kubenswrapper[4824]: I1209 09:47:39.618310 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:39 crc kubenswrapper[4824]: I1209 09:47:39.618374 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:39 crc kubenswrapper[4824]: I1209 09:47:39.618392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:39 crc kubenswrapper[4824]: I1209 09:47:39.625634 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.013972 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.014945 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.014991 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.015005 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.397313 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.399049 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.399095 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.399111 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.399141 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 09:47:40 crc kubenswrapper[4824]: E1209 09:47:40.403930 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.785288 4824 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.858812 4824 apiserver.go:52] "Watching apiserver" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.861944 4824 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.862250 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.862559 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.862989 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:47:40 crc kubenswrapper[4824]: E1209 09:47:40.863043 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.863100 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:47:40 crc kubenswrapper[4824]: E1209 09:47:40.863122 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.863159 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.863195 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.863463 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:47:40 crc kubenswrapper[4824]: E1209 09:47:40.863551 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.864904 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.865333 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.865947 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.866020 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.866570 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.866910 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.867179 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.867373 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.867326 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.896691 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.909082 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.920919 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.935416 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.947880 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.961155 4824 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.963185 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 09:47:40 crc kubenswrapper[4824]: I1209 09:47:40.978305 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 09:47:41 crc kubenswrapper[4824]: I1209 09:47:41.532277 4824 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.828092 4824 trace.go:236] Trace[1874386999]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Dec-2025 09:47:27.972) (total time: 14855ms): Dec 09 09:47:42 crc kubenswrapper[4824]: Trace[1874386999]: ---"Objects listed" error: 14855ms (09:47:42.827) Dec 09 09:47:42 crc kubenswrapper[4824]: Trace[1874386999]: [14.8553702s] [14.8553702s] END Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.828123 4824 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.828469 4824 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.856534 4824 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.876542 4824 csr.go:261] certificate signing request csr-x4g5l is approved, waiting to be issued Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.890316 4824 csr.go:257] certificate signing request csr-x4g5l is issued Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.909536 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.909600 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.909702 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:47:42 crc kubenswrapper[4824]: E1209 09:47:42.909841 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:47:42 crc kubenswrapper[4824]: E1209 09:47:42.909938 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:47:42 crc kubenswrapper[4824]: E1209 09:47:42.910100 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929104 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929146 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929170 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929188 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929244 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929278 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929294 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929310 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929327 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929344 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929361 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929378 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929398 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929423 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929465 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929482 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929501 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929517 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929532 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929548 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929564 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929602 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929620 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929601 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929635 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929780 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929754 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929773 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929867 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929896 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929935 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929971 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929992 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.930014 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.930190 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.930217 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.930237 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.930262 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.930288 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.930306 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.930332 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.930354 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931150 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931537 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931567 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931588 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931609 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931632 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931654 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931682 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931711 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931734 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931762 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931800 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931823 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931847 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931871 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931894 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931914 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931932 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931954 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931976 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931994 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932022 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932045 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932067 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932086 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932108 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932130 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932149 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932192 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932213 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932232 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932250 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932277 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932301 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932319 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932339 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932358 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932604 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932633 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932655 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932695 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932717 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932734 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932753 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932773 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932817 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932836 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932925 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932946 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932967 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933012 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933043 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933071 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933109 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933137 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933170 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933198 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933225 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933249 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933274 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933296 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933321 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933345 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933370 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933388 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933408 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933426 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933444 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933464 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933483 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933501 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933529 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933560 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933582 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933605 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933633 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933653 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933671 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933688 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933707 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933728 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933745 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.933763 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.934551 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.934585 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.934605 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.934622 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.934641 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.934658 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.934676 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.934693 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.934712 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.934729 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.934747 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.934766 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.934801 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.929850 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.930057 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.930094 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.930298 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.930322 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.935477 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.937271 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.938004 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.938041 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.940900 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.940954 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.941114 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.930355 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.930408 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.930541 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.930638 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.930643 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.930879 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.930962 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931028 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931114 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931316 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931591 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931834 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931857 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.931997 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932087 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932197 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932227 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932339 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932385 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932652 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.932890 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.934156 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.934213 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.934898 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.941148 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.941217 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.941249 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.941398 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.938844 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.941290 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.941526 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.941570 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.941721 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.941725 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.941755 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.941923 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.941982 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.941998 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.942179 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.942259 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.942590 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.942683 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.942717 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943098 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943123 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943143 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943161 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943177 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943204 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943224 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943244 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943283 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943302 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943319 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943337 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943352 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943372 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943389 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943405 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943421 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943437 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943454 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943470 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943486 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943504 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943525 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943541 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943557 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943576 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943592 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943608 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943623 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943638 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943654 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943673 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943689 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943707 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943722 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943739 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943755 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943773 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943803 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943821 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943839 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943856 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943870 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943886 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943901 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943930 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943947 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943963 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943979 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.943994 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.944010 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.944026 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.944041 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.944067 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.944087 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.944104 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.944129 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.944187 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.944212 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.944231 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.944250 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.944268 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.944286 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.944306 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.944327 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.944343 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.944361 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.944378 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.944408 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.944421 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.944428 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.944489 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.946905 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.947265 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.947514 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.948209 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.948454 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.948684 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.949005 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.949428 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.949976 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.950013 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.950100 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.950208 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.950214 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.950616 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.950915 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.951168 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.952089 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.952161 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.952210 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.952238 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.944519 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.952327 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.952567 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.952677 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.952765 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.952829 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.952843 4824 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.952857 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.952871 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.952887 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.952900 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.952912 4824 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.952928 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.952942 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.952953 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.952963 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.952978 4824 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.952992 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953003 4824 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953014 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953028 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953039 4824 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953049 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953077 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953095 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953109 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953123 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953144 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953157 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953170 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953181 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953196 4824 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953206 4824 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953216 4824 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953226 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953241 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953255 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953267 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953281 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953291 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953300 4824 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953310 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953323 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953334 4824 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953344 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953354 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953367 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953377 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953387 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953397 4824 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953410 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953421 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953431 4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953444 4824 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953454 4824 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953464 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953474 4824 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953487 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953497 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953509 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953519 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953538 4824 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953547 4824 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953557 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953560 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953559 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953753 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953570 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953949 4824 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953976 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953988 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.953994 4824 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.954079 4824 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.954084 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.954097 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.954098 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.954240 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.954283 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.954653 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.954677 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.955014 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.955401 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.955433 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.955760 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.955860 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.955794 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.956060 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.956067 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.956006 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.956168 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.956338 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.956332 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.956374 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.956618 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.956718 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.956884 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.957365 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.957427 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.957937 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.958014 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.958459 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.958955 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.959026 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.959048 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.959064 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.959110 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.959136 4824 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.959178 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.959152 4824 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.959366 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.959504 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.959830 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.960056 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: E1209 09:47:42.960201 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:47:43.460145916 +0000 UTC m=+19.794650583 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.960215 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.960215 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.960394 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.961075 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.960417 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.960466 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.960609 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.960762 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.960831 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.960674 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.961112 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.961803 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.961945 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.962309 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.962437 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.963409 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.964168 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.964359 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.959969 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.964951 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.968455 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.969408 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.969432 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.970460 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.971384 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.971427 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.970810 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.971601 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.971617 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.971433 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.972332 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.972976 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.973089 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.973254 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.973674 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.973615 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.974041 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.973334 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.974148 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.974192 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.974718 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.974845 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.974861 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.974876 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.974861 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.975202 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.975551 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.975030 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.975414 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.975429 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.975936 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.975966 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.978678 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.978835 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.978838 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: E1209 09:47:42.978925 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 09:47:42 crc kubenswrapper[4824]: E1209 09:47:42.979029 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 09:47:43.479002448 +0000 UTC m=+19.813507245 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.979065 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.979110 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.979190 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.979318 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.979719 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.979775 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.979767 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.979912 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.979947 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: E1209 09:47:42.980090 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 09:47:42 crc kubenswrapper[4824]: E1209 09:47:42.980196 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 09:47:43.480142343 +0000 UTC m=+19.814647200 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.981301 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.981381 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.981568 4824 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.983639 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.987793 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.994109 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.995318 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.995385 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.995402 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.996302 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.997145 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.997659 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.997843 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.998092 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.999498 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:42 crc kubenswrapper[4824]: E1209 09:47:42.999595 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 09:47:42 crc kubenswrapper[4824]: E1209 09:47:42.999646 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 09:47:42 crc kubenswrapper[4824]: E1209 09:47:42.999659 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:47:42 crc kubenswrapper[4824]: E1209 09:47:42.999718 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 09:47:43.499695656 +0000 UTC m=+19.834200323 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:47:42 crc kubenswrapper[4824]: I1209 09:47:42.999823 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.005113 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:47:43 crc kubenswrapper[4824]: E1209 09:47:43.005374 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 09:47:43 crc kubenswrapper[4824]: E1209 09:47:43.005412 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 09:47:43 crc kubenswrapper[4824]: E1209 09:47:43.005430 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:47:43 crc kubenswrapper[4824]: E1209 09:47:43.005502 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 09:47:43.505471967 +0000 UTC m=+19.839976824 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.006765 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.006981 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.010072 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.013725 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.014219 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.017565 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.030982 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.038576 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.039830 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:41748->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.039921 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:41748->192.168.126.11:17697: read: connection reset by peer" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.040415 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.040477 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.040997 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.041041 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.044014 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065359 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065429 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065497 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065545 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065557 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065568 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065607 4824 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065629 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065652 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065661 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065670 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065678 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065688 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065699 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065709 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065719 4824 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065727 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065737 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065745 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065753 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065761 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065769 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065777 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065816 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065829 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065837 4824 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065818 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065845 4824 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065943 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065967 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.065987 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066004 4824 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066021 4824 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066037 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066053 4824 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066069 4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066085 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066099 4824 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066114 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066130 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066146 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066161 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066177 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066192 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066206 4824 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066225 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066241 4824 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066256 4824 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066272 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066286 4824 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066301 4824 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066316 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066331 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066346 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066361 4824 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066377 4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066391 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066408 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066427 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066442 4824 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066468 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066482 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066499 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066513 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066528 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066542 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066557 4824 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066570 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066587 4824 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066600 4824 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066614 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066628 4824 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066641 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066656 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066670 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066683 4824 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066699 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066712 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066725 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066740 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066754 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066767 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066804 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066820 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066834 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066848 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066865 4824 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066879 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066913 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066927 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066940 4824 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066955 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066969 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066982 4824 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.066999 4824 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067014 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067028 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067047 4824 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067062 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067077 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067091 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067105 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067120 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067134 4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067149 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067162 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067176 4824 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067190 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067204 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067217 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067231 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067244 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067258 4824 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067272 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067286 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067299 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067312 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067326 4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067338 4824 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067352 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067368 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.067381 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.267530 4824 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.281874 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.289503 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.296868 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 09:47:43 crc kubenswrapper[4824]: W1209 09:47:43.298563 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-c6d2cdbc884bf99cd78fc34e5cf574bd1127339f822a5099213c1f02cba1d5a3 WatchSource:0}: Error finding container c6d2cdbc884bf99cd78fc34e5cf574bd1127339f822a5099213c1f02cba1d5a3: Status 404 returned error can't find the container with id c6d2cdbc884bf99cd78fc34e5cf574bd1127339f822a5099213c1f02cba1d5a3 Dec 09 09:47:43 crc kubenswrapper[4824]: W1209 09:47:43.308173 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-a2fe59a2e5663e7a63669fc8cb2ce15ac2e56cd926bfbbd23d6b6f5ccce5c323 WatchSource:0}: Error finding container a2fe59a2e5663e7a63669fc8cb2ce15ac2e56cd926bfbbd23d6b6f5ccce5c323: Status 404 returned error can't find the container with id a2fe59a2e5663e7a63669fc8cb2ce15ac2e56cd926bfbbd23d6b6f5ccce5c323 Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.409458 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.414301 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.417848 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.420343 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.430902 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.440408 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.449634 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.459455 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.471114 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:47:43 crc kubenswrapper[4824]: E1209 09:47:43.471315 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:47:44.471290978 +0000 UTC m=+20.805795645 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.471799 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.483946 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.496309 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.508895 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.520149 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.532834 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.546208 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.556890 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 09:47:43 crc kubenswrapper[4824]: E1209 09:47:43.572931 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 09:47:43 crc kubenswrapper[4824]: E1209 09:47:43.572978 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 09:47:43 crc kubenswrapper[4824]: E1209 09:47:43.572994 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:47:43 crc kubenswrapper[4824]: E1209 09:47:43.573078 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 09:47:44.573058529 +0000 UTC m=+20.907563196 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.572717 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.573178 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.573215 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.573246 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:47:43 crc kubenswrapper[4824]: E1209 09:47:43.573330 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 09:47:43 crc kubenswrapper[4824]: E1209 09:47:43.573360 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 09:47:44.573352038 +0000 UTC m=+20.907856705 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 09:47:43 crc kubenswrapper[4824]: E1209 09:47:43.573467 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 09:47:43 crc kubenswrapper[4824]: E1209 09:47:43.573481 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 09:47:43 crc kubenswrapper[4824]: E1209 09:47:43.573489 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:47:43 crc kubenswrapper[4824]: E1209 09:47:43.573475 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 09:47:43 crc kubenswrapper[4824]: E1209 09:47:43.573519 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 09:47:44.573510863 +0000 UTC m=+20.908015540 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:47:43 crc kubenswrapper[4824]: E1209 09:47:43.573630 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 09:47:44.573609876 +0000 UTC m=+20.908114543 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.796828 4824 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 09 09:47:43 crc kubenswrapper[4824]: W1209 09:47:43.797269 4824 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.CSIDriver ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Dec 09 09:47:43 crc kubenswrapper[4824]: E1209 09:47:43.797285 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Post \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases?timeout=10s\": read tcp 38.102.83.128:53412->38.102.83.128:6443: use of closed network connection" interval="6.4s" Dec 09 09:47:43 crc kubenswrapper[4824]: W1209 09:47:43.797408 4824 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.Service ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Dec 09 09:47:43 crc kubenswrapper[4824]: E1209 09:47:43.798091 4824 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/events\": read tcp 38.102.83.128:53412->38.102.83.128:6443: use of closed network connection" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.187f8306e50e163e openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-09 09:47:24.58847187 +0000 UTC m=+0.922976537,LastTimestamp:2025-12-09 09:47:24.58847187 +0000 UTC m=+0.922976537,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.891669 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-09 09:42:42 +0000 UTC, rotation deadline is 2026-09-04 19:05:02.847430189 +0000 UTC Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.891820 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6465h17m18.955613933s for next certificate rotation Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.914721 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.915370 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.917030 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.917862 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.919040 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.919725 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.920485 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.921689 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.922589 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.924541 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.925224 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.926625 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.927329 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.928006 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.930448 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.931111 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.932360 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.932887 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.933621 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.938241 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.939696 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.940382 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.941699 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.942309 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.943730 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.944353 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.945181 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.946508 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.947107 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.948308 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.948952 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.950032 4824 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.950157 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.952104 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.953175 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.953718 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.955484 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.956293 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.957318 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.958119 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.959382 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.960005 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.961297 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.962105 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.963292 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.963898 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.965032 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.965196 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.965701 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.967150 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.967656 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.968616 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.969207 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.970355 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.971261 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.971799 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 09 09:47:43 crc kubenswrapper[4824]: I1209 09:47:43.981632 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.000419 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.016632 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.035220 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.037614 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f"} Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.037709 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928"} Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.037728 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a2fe59a2e5663e7a63669fc8cb2ce15ac2e56cd926bfbbd23d6b6f5ccce5c323"} Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.038896 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357"} Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.038968 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"c6d2cdbc884bf99cd78fc34e5cf574bd1127339f822a5099213c1f02cba1d5a3"} Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.041440 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.043712 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042" exitCode=255 Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.043759 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042"} Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.046378 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"bcdcd24ecbbd55bc859f1c86c2fb931fa452760049161134eb4e46d7084ad4d8"} Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.054432 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.055557 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.056053 4824 scope.go:117] "RemoveContainer" containerID="723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.072873 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.089270 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.107405 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.120323 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.137063 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.165462 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.194129 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.225764 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-zhqgl"] Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.226101 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-zhqgl" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.228881 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 09 09:47:44 crc kubenswrapper[4824]: W1209 09:47:44.228937 4824 reflector.go:561] object-"openshift-dns"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Dec 09 09:47:44 crc kubenswrapper[4824]: E1209 09:47:44.228992 4824 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.230425 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.248812 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.272762 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.288927 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.312369 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.326255 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.339153 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.378260 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.380526 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf-hosts-file\") pod \"node-resolver-zhqgl\" (UID: \"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\") " pod="openshift-dns/node-resolver-zhqgl" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.380660 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk9bt\" (UniqueName: \"kubernetes.io/projected/1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf-kube-api-access-lk9bt\") pod \"node-resolver-zhqgl\" (UID: \"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\") " pod="openshift-dns/node-resolver-zhqgl" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.394826 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.412052 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.481390 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.481481 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf-hosts-file\") pod \"node-resolver-zhqgl\" (UID: \"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\") " pod="openshift-dns/node-resolver-zhqgl" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.481510 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk9bt\" (UniqueName: \"kubernetes.io/projected/1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf-kube-api-access-lk9bt\") pod \"node-resolver-zhqgl\" (UID: \"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\") " pod="openshift-dns/node-resolver-zhqgl" Dec 09 09:47:44 crc kubenswrapper[4824]: E1209 09:47:44.481571 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:47:46.481535852 +0000 UTC m=+22.816040509 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.481656 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf-hosts-file\") pod \"node-resolver-zhqgl\" (UID: \"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\") " pod="openshift-dns/node-resolver-zhqgl" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.582924 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.583237 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.583335 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.583414 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:47:44 crc kubenswrapper[4824]: E1209 09:47:44.583124 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 09:47:44 crc kubenswrapper[4824]: E1209 09:47:44.583574 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 09:47:44 crc kubenswrapper[4824]: E1209 09:47:44.583596 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 09:47:46.583573542 +0000 UTC m=+22.918078209 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 09:47:44 crc kubenswrapper[4824]: E1209 09:47:44.583507 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 09:47:44 crc kubenswrapper[4824]: E1209 09:47:44.583738 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 09:47:44 crc kubenswrapper[4824]: E1209 09:47:44.583753 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:47:44 crc kubenswrapper[4824]: E1209 09:47:44.583391 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 09:47:44 crc kubenswrapper[4824]: E1209 09:47:44.583838 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 09:47:44 crc kubenswrapper[4824]: E1209 09:47:44.583847 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:47:44 crc kubenswrapper[4824]: E1209 09:47:44.583711 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 09:47:46.583687156 +0000 UTC m=+22.918191813 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 09:47:44 crc kubenswrapper[4824]: E1209 09:47:44.583901 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 09:47:46.583878762 +0000 UTC m=+22.918383479 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:47:44 crc kubenswrapper[4824]: E1209 09:47:44.583924 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 09:47:46.583909073 +0000 UTC m=+22.918413840 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.910242 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.910295 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:47:44 crc kubenswrapper[4824]: E1209 09:47:44.911453 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:47:44 crc kubenswrapper[4824]: I1209 09:47:44.910388 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:47:44 crc kubenswrapper[4824]: E1209 09:47:44.911705 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:47:44 crc kubenswrapper[4824]: E1209 09:47:44.911542 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.050743 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.052387 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598"} Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.052750 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.106437 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.126520 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.144764 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk9bt\" (UniqueName: \"kubernetes.io/projected/1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf-kube-api-access-lk9bt\") pod \"node-resolver-zhqgl\" (UID: \"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\") " pod="openshift-dns/node-resolver-zhqgl" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.149553 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-dth8x"] Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.149939 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-pvksg"] Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.150188 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.150548 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.154212 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.154891 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.154930 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.155633 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.158205 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.158359 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.158483 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.158669 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.160647 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.161644 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.164158 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-w8sjn"] Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.164755 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.177104 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.177114 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.183491 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.199938 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.219748 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.233003 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.246766 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.262409 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.281721 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.289733 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-system-cni-dir\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.289824 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9e41bcf8-bacf-4eb1-a391-bbd86adedb32-cni-binary-copy\") pod \"multus-additional-cni-plugins-w8sjn\" (UID: \"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\") " pod="openshift-multus/multus-additional-cni-plugins-w8sjn" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.289858 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9e41bcf8-bacf-4eb1-a391-bbd86adedb32-tuning-conf-dir\") pod \"multus-additional-cni-plugins-w8sjn\" (UID: \"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\") " pod="openshift-multus/multus-additional-cni-plugins-w8sjn" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.289888 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-hostroot\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.289912 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9e41bcf8-bacf-4eb1-a391-bbd86adedb32-os-release\") pod \"multus-additional-cni-plugins-w8sjn\" (UID: \"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\") " pod="openshift-multus/multus-additional-cni-plugins-w8sjn" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.289935 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/74aa3cd7-0abc-4fc9-9697-e23b693a7611-multus-daemon-config\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.289960 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcwxt\" (UniqueName: \"kubernetes.io/projected/9e41bcf8-bacf-4eb1-a391-bbd86adedb32-kube-api-access-qcwxt\") pod \"multus-additional-cni-plugins-w8sjn\" (UID: \"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\") " pod="openshift-multus/multus-additional-cni-plugins-w8sjn" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.289986 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-host-run-k8s-cni-cncf-io\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.290011 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wtmd\" (UniqueName: \"kubernetes.io/projected/ee3c86bf-f246-42cb-adb6-900e3d735166-kube-api-access-5wtmd\") pod \"machine-config-daemon-dth8x\" (UID: \"ee3c86bf-f246-42cb-adb6-900e3d735166\") " pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.290037 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-multus-socket-dir-parent\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.290063 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-host-var-lib-cni-bin\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.290100 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-cnibin\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.290125 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-multus-cni-dir\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.290149 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-etc-kubernetes\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.290172 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-host-run-netns\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.290206 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-host-run-multus-certs\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.290230 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-os-release\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.290254 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs2pc\" (UniqueName: \"kubernetes.io/projected/74aa3cd7-0abc-4fc9-9697-e23b693a7611-kube-api-access-bs2pc\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.290289 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ee3c86bf-f246-42cb-adb6-900e3d735166-mcd-auth-proxy-config\") pod \"machine-config-daemon-dth8x\" (UID: \"ee3c86bf-f246-42cb-adb6-900e3d735166\") " pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.290310 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-host-var-lib-kubelet\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.290332 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ee3c86bf-f246-42cb-adb6-900e3d735166-rootfs\") pod \"machine-config-daemon-dth8x\" (UID: \"ee3c86bf-f246-42cb-adb6-900e3d735166\") " pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.290356 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-host-var-lib-cni-multus\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.290378 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ee3c86bf-f246-42cb-adb6-900e3d735166-proxy-tls\") pod \"machine-config-daemon-dth8x\" (UID: \"ee3c86bf-f246-42cb-adb6-900e3d735166\") " pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.290406 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-multus-conf-dir\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.290431 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9e41bcf8-bacf-4eb1-a391-bbd86adedb32-system-cni-dir\") pod \"multus-additional-cni-plugins-w8sjn\" (UID: \"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\") " pod="openshift-multus/multus-additional-cni-plugins-w8sjn" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.290458 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9e41bcf8-bacf-4eb1-a391-bbd86adedb32-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-w8sjn\" (UID: \"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\") " pod="openshift-multus/multus-additional-cni-plugins-w8sjn" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.290493 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/74aa3cd7-0abc-4fc9-9697-e23b693a7611-cni-binary-copy\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.290514 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9e41bcf8-bacf-4eb1-a391-bbd86adedb32-cnibin\") pod \"multus-additional-cni-plugins-w8sjn\" (UID: \"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\") " pod="openshift-multus/multus-additional-cni-plugins-w8sjn" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.295772 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.311824 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.324044 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.338588 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.355298 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.368941 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.386445 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.391054 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-host-run-netns\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.391087 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-host-run-multus-certs\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.391104 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-etc-kubernetes\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.391121 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-os-release\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.391141 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs2pc\" (UniqueName: \"kubernetes.io/projected/74aa3cd7-0abc-4fc9-9697-e23b693a7611-kube-api-access-bs2pc\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.391168 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-host-var-lib-kubelet\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.391187 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ee3c86bf-f246-42cb-adb6-900e3d735166-rootfs\") pod \"machine-config-daemon-dth8x\" (UID: \"ee3c86bf-f246-42cb-adb6-900e3d735166\") " pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.391203 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ee3c86bf-f246-42cb-adb6-900e3d735166-mcd-auth-proxy-config\") pod \"machine-config-daemon-dth8x\" (UID: \"ee3c86bf-f246-42cb-adb6-900e3d735166\") " pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.391221 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-host-var-lib-cni-multus\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.391237 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ee3c86bf-f246-42cb-adb6-900e3d735166-proxy-tls\") pod \"machine-config-daemon-dth8x\" (UID: \"ee3c86bf-f246-42cb-adb6-900e3d735166\") " pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.391246 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-host-run-netns\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.391294 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-multus-conf-dir\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.391252 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-multus-conf-dir\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.391324 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-host-var-lib-cni-multus\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.391333 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9e41bcf8-bacf-4eb1-a391-bbd86adedb32-system-cni-dir\") pod \"multus-additional-cni-plugins-w8sjn\" (UID: \"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\") " pod="openshift-multus/multus-additional-cni-plugins-w8sjn" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.391354 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9e41bcf8-bacf-4eb1-a391-bbd86adedb32-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-w8sjn\" (UID: \"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\") " pod="openshift-multus/multus-additional-cni-plugins-w8sjn" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.391239 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-host-run-multus-certs\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.391378 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ee3c86bf-f246-42cb-adb6-900e3d735166-rootfs\") pod \"machine-config-daemon-dth8x\" (UID: \"ee3c86bf-f246-42cb-adb6-900e3d735166\") " pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.391385 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9e41bcf8-bacf-4eb1-a391-bbd86adedb32-system-cni-dir\") pod \"multus-additional-cni-plugins-w8sjn\" (UID: \"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\") " pod="openshift-multus/multus-additional-cni-plugins-w8sjn" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.391423 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-host-var-lib-kubelet\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.391355 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-etc-kubernetes\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.391590 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-os-release\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.392239 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ee3c86bf-f246-42cb-adb6-900e3d735166-mcd-auth-proxy-config\") pod \"machine-config-daemon-dth8x\" (UID: \"ee3c86bf-f246-42cb-adb6-900e3d735166\") " pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.392321 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9e41bcf8-bacf-4eb1-a391-bbd86adedb32-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-w8sjn\" (UID: \"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\") " pod="openshift-multus/multus-additional-cni-plugins-w8sjn" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.392374 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/74aa3cd7-0abc-4fc9-9697-e23b693a7611-cni-binary-copy\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.392441 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9e41bcf8-bacf-4eb1-a391-bbd86adedb32-cnibin\") pod \"multus-additional-cni-plugins-w8sjn\" (UID: \"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\") " pod="openshift-multus/multus-additional-cni-plugins-w8sjn" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.392481 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-system-cni-dir\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.392490 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9e41bcf8-bacf-4eb1-a391-bbd86adedb32-cnibin\") pod \"multus-additional-cni-plugins-w8sjn\" (UID: \"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\") " pod="openshift-multus/multus-additional-cni-plugins-w8sjn" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.392502 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9e41bcf8-bacf-4eb1-a391-bbd86adedb32-cni-binary-copy\") pod \"multus-additional-cni-plugins-w8sjn\" (UID: \"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\") " pod="openshift-multus/multus-additional-cni-plugins-w8sjn" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.392551 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9e41bcf8-bacf-4eb1-a391-bbd86adedb32-tuning-conf-dir\") pod \"multus-additional-cni-plugins-w8sjn\" (UID: \"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\") " pod="openshift-multus/multus-additional-cni-plugins-w8sjn" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.392550 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-system-cni-dir\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.392581 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-hostroot\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.392610 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9e41bcf8-bacf-4eb1-a391-bbd86adedb32-os-release\") pod \"multus-additional-cni-plugins-w8sjn\" (UID: \"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\") " pod="openshift-multus/multus-additional-cni-plugins-w8sjn" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.392630 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-hostroot\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.392635 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-host-run-k8s-cni-cncf-io\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.392661 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/74aa3cd7-0abc-4fc9-9697-e23b693a7611-multus-daemon-config\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.392679 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9e41bcf8-bacf-4eb1-a391-bbd86adedb32-os-release\") pod \"multus-additional-cni-plugins-w8sjn\" (UID: \"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\") " pod="openshift-multus/multus-additional-cni-plugins-w8sjn" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.392689 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcwxt\" (UniqueName: \"kubernetes.io/projected/9e41bcf8-bacf-4eb1-a391-bbd86adedb32-kube-api-access-qcwxt\") pod \"multus-additional-cni-plugins-w8sjn\" (UID: \"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\") " pod="openshift-multus/multus-additional-cni-plugins-w8sjn" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.392730 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-host-run-k8s-cni-cncf-io\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.392763 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wtmd\" (UniqueName: \"kubernetes.io/projected/ee3c86bf-f246-42cb-adb6-900e3d735166-kube-api-access-5wtmd\") pod \"machine-config-daemon-dth8x\" (UID: \"ee3c86bf-f246-42cb-adb6-900e3d735166\") " pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.392812 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-multus-socket-dir-parent\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.392854 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-cnibin\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.392871 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-host-var-lib-cni-bin\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.392893 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-multus-cni-dir\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.392893 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9e41bcf8-bacf-4eb1-a391-bbd86adedb32-tuning-conf-dir\") pod \"multus-additional-cni-plugins-w8sjn\" (UID: \"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\") " pod="openshift-multus/multus-additional-cni-plugins-w8sjn" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.392967 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-multus-socket-dir-parent\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.393006 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-cnibin\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.393014 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-host-var-lib-cni-bin\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.393182 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/74aa3cd7-0abc-4fc9-9697-e23b693a7611-multus-cni-dir\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.393319 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/74aa3cd7-0abc-4fc9-9697-e23b693a7611-multus-daemon-config\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.393632 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/74aa3cd7-0abc-4fc9-9697-e23b693a7611-cni-binary-copy\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.397534 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9e41bcf8-bacf-4eb1-a391-bbd86adedb32-cni-binary-copy\") pod \"multus-additional-cni-plugins-w8sjn\" (UID: \"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\") " pod="openshift-multus/multus-additional-cni-plugins-w8sjn" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.405252 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ee3c86bf-f246-42cb-adb6-900e3d735166-proxy-tls\") pod \"machine-config-daemon-dth8x\" (UID: \"ee3c86bf-f246-42cb-adb6-900e3d735166\") " pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.406063 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.409021 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs2pc\" (UniqueName: \"kubernetes.io/projected/74aa3cd7-0abc-4fc9-9697-e23b693a7611-kube-api-access-bs2pc\") pod \"multus-pvksg\" (UID: \"74aa3cd7-0abc-4fc9-9697-e23b693a7611\") " pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.409212 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcwxt\" (UniqueName: \"kubernetes.io/projected/9e41bcf8-bacf-4eb1-a391-bbd86adedb32-kube-api-access-qcwxt\") pod \"multus-additional-cni-plugins-w8sjn\" (UID: \"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\") " pod="openshift-multus/multus-additional-cni-plugins-w8sjn" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.415937 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wtmd\" (UniqueName: \"kubernetes.io/projected/ee3c86bf-f246-42cb-adb6-900e3d735166-kube-api-access-5wtmd\") pod \"machine-config-daemon-dth8x\" (UID: \"ee3c86bf-f246-42cb-adb6-900e3d735166\") " pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.421952 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.434036 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.439246 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-zhqgl" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.448434 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.462858 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.464971 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-pvksg" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.472700 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.477882 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.479471 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.529301 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vvxck"] Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.530148 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.532328 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.533396 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.533637 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.533716 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.533957 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.534010 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.534297 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.547907 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: W1209 09:47:45.554408 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c4d3c84_c8c8_4c5b_8e15_95ab59ea77cf.slice/crio-fd768132206a5049687fae9b8348a73e1b48b52726b27ea120b245c77b5e22bd WatchSource:0}: Error finding container fd768132206a5049687fae9b8348a73e1b48b52726b27ea120b245c77b5e22bd: Status 404 returned error can't find the container with id fd768132206a5049687fae9b8348a73e1b48b52726b27ea120b245c77b5e22bd Dec 09 09:47:45 crc kubenswrapper[4824]: W1209 09:47:45.555288 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e41bcf8_bacf_4eb1_a391_bbd86adedb32.slice/crio-2bcb9c9ceeea06b5a5dc7738eb5d36700041be802283454e032882a91e15df77 WatchSource:0}: Error finding container 2bcb9c9ceeea06b5a5dc7738eb5d36700041be802283454e032882a91e15df77: Status 404 returned error can't find the container with id 2bcb9c9ceeea06b5a5dc7738eb5d36700041be802283454e032882a91e15df77 Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.562286 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.577208 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.594548 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.609749 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.624672 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.640105 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.653854 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.667956 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.681634 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.695133 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.696358 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-run-systemd\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.696400 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrd4m\" (UniqueName: \"kubernetes.io/projected/13620d58-6925-48c7-bbc7-f34e92922480-kube-api-access-zrd4m\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.697001 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-slash\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.697154 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.697188 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/13620d58-6925-48c7-bbc7-f34e92922480-ovnkube-config\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.697244 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-run-ovn\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.697307 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-cni-bin\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.697533 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-etc-openvswitch\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.697699 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-run-netns\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.697763 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-cni-netd\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.697811 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-systemd-units\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.697830 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/13620d58-6925-48c7-bbc7-f34e92922480-ovnkube-script-lib\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.697855 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-kubelet\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.697873 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-log-socket\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.697892 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-run-ovn-kubernetes\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.697924 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/13620d58-6925-48c7-bbc7-f34e92922480-ovn-node-metrics-cert\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.697968 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-node-log\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.698006 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-var-lib-openvswitch\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.698030 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-run-openvswitch\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.698050 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/13620d58-6925-48c7-bbc7-f34e92922480-env-overrides\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.711051 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.735255 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.798702 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-run-ovn-kubernetes\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.798749 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-kubelet\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.798773 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-log-socket\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.798825 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/13620d58-6925-48c7-bbc7-f34e92922480-ovn-node-metrics-cert\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.798836 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-run-ovn-kubernetes\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.798860 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-node-log\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.798879 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-var-lib-openvswitch\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.798897 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-run-openvswitch\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.798915 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/13620d58-6925-48c7-bbc7-f34e92922480-env-overrides\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.798935 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-run-systemd\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.798954 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrd4m\" (UniqueName: \"kubernetes.io/projected/13620d58-6925-48c7-bbc7-f34e92922480-kube-api-access-zrd4m\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.798958 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-node-log\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.798983 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-slash\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.799002 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.799020 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/13620d58-6925-48c7-bbc7-f34e92922480-ovnkube-config\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.799040 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-run-ovn\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.799055 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-cni-bin\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.799077 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-etc-openvswitch\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.799052 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-kubelet\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.799139 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-run-netns\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.799063 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-log-socket\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.799162 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-run-openvswitch\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.799136 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-var-lib-openvswitch\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.799185 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.799241 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-slash\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.799094 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-run-netns\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.799268 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-cni-bin\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.799274 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-cni-netd\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.799293 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-cni-netd\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.799374 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-systemd-units\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.799476 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-run-systemd\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.799512 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-run-ovn\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.799537 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-etc-openvswitch\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.799569 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-systemd-units\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.799609 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/13620d58-6925-48c7-bbc7-f34e92922480-ovnkube-script-lib\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.799655 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/13620d58-6925-48c7-bbc7-f34e92922480-env-overrides\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.799720 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/13620d58-6925-48c7-bbc7-f34e92922480-ovnkube-config\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.800534 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/13620d58-6925-48c7-bbc7-f34e92922480-ovnkube-script-lib\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.802343 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/13620d58-6925-48c7-bbc7-f34e92922480-ovn-node-metrics-cert\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.816167 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrd4m\" (UniqueName: \"kubernetes.io/projected/13620d58-6925-48c7-bbc7-f34e92922480-kube-api-access-zrd4m\") pod \"ovnkube-node-vvxck\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.845941 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.850408 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 09 09:47:45 crc kubenswrapper[4824]: W1209 09:47:45.858261 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13620d58_6925_48c7_bbc7_f34e92922480.slice/crio-b0d26dac1a3a6325cac749ee337191fb4c11cb083c7dfc4a4afc10c4310fdbb3 WatchSource:0}: Error finding container b0d26dac1a3a6325cac749ee337191fb4c11cb083c7dfc4a4afc10c4310fdbb3: Status 404 returned error can't find the container with id b0d26dac1a3a6325cac749ee337191fb4c11cb083c7dfc4a4afc10c4310fdbb3 Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.866683 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.867750 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.872356 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.879426 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.893141 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.943912 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.961967 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.974515 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:45 crc kubenswrapper[4824]: I1209 09:47:45.990070 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:45Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.008577 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.020945 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.030796 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.042492 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.061338 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde"} Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.062162 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.063461 4824 generic.go:334] "Generic (PLEG): container finished" podID="13620d58-6925-48c7-bbc7-f34e92922480" containerID="21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339" exitCode=0 Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.063526 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerDied","Data":"21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339"} Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.063547 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerStarted","Data":"b0d26dac1a3a6325cac749ee337191fb4c11cb083c7dfc4a4afc10c4310fdbb3"} Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.065235 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pvksg" event={"ID":"74aa3cd7-0abc-4fc9-9697-e23b693a7611","Type":"ContainerStarted","Data":"cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de"} Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.065282 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pvksg" event={"ID":"74aa3cd7-0abc-4fc9-9697-e23b693a7611","Type":"ContainerStarted","Data":"a7806ee3b974d1a25c34bdd9f834f66efc5e65ecf6e5e217e045806c5186ccdb"} Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.067522 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-zhqgl" event={"ID":"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf","Type":"ContainerStarted","Data":"4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5"} Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.067585 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-zhqgl" event={"ID":"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf","Type":"ContainerStarted","Data":"fd768132206a5049687fae9b8348a73e1b48b52726b27ea120b245c77b5e22bd"} Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.069548 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" event={"ID":"9e41bcf8-bacf-4eb1-a391-bbd86adedb32","Type":"ContainerStarted","Data":"660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b"} Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.069581 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" event={"ID":"9e41bcf8-bacf-4eb1-a391-bbd86adedb32","Type":"ContainerStarted","Data":"2bcb9c9ceeea06b5a5dc7738eb5d36700041be802283454e032882a91e15df77"} Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.072856 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerStarted","Data":"314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671"} Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.072911 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerStarted","Data":"c1bd609d5296a866f9eb4e0004262fbdaca7f5ae4a3c8eea79dfb4082f13d62c"} Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.084651 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: E1209 09:47:46.086006 4824 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.107042 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.123307 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.139294 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.153685 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.166657 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.181496 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.197742 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.218841 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.235614 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.252051 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.264867 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.279143 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.303300 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.319320 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.509449 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:47:46 crc kubenswrapper[4824]: E1209 09:47:46.509734 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:47:50.509692052 +0000 UTC m=+26.844196719 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.610556 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:47:46 crc kubenswrapper[4824]: E1209 09:47:46.610876 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 09:47:46 crc kubenswrapper[4824]: E1209 09:47:46.611032 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 09:47:46 crc kubenswrapper[4824]: E1209 09:47:46.611047 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 09:47:46 crc kubenswrapper[4824]: E1209 09:47:46.611095 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 09:47:50.611079772 +0000 UTC m=+26.945584439 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.610989 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:47:46 crc kubenswrapper[4824]: E1209 09:47:46.611048 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:47:46 crc kubenswrapper[4824]: E1209 09:47:46.611196 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 09:47:46 crc kubenswrapper[4824]: E1209 09:47:46.611208 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 09:47:46 crc kubenswrapper[4824]: E1209 09:47:46.611216 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 09:47:50.611183635 +0000 UTC m=+26.945688302 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:47:46 crc kubenswrapper[4824]: E1209 09:47:46.611219 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.611127 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:47:46 crc kubenswrapper[4824]: E1209 09:47:46.611256 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 09:47:50.611249657 +0000 UTC m=+26.945754324 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.611294 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:47:46 crc kubenswrapper[4824]: E1209 09:47:46.611361 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 09:47:46 crc kubenswrapper[4824]: E1209 09:47:46.611392 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 09:47:50.611384732 +0000 UTC m=+26.945889399 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.805024 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.807111 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.807157 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.807165 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.807253 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.813187 4824 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.813402 4824 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.814418 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.814466 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.814479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.814501 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.814515 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:46Z","lastTransitionTime":"2025-12-09T09:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:46 crc kubenswrapper[4824]: E1209 09:47:46.833325 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.836818 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.836853 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.836865 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.836882 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.836892 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:46Z","lastTransitionTime":"2025-12-09T09:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:46 crc kubenswrapper[4824]: E1209 09:47:46.852769 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.855902 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.855941 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.855954 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.855970 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.855980 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:46Z","lastTransitionTime":"2025-12-09T09:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:46 crc kubenswrapper[4824]: E1209 09:47:46.867418 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.871066 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.871099 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.871109 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.871123 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.871134 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:46Z","lastTransitionTime":"2025-12-09T09:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:46 crc kubenswrapper[4824]: E1209 09:47:46.883060 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.886685 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.886725 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.886743 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.886764 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.886776 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:46Z","lastTransitionTime":"2025-12-09T09:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:46 crc kubenswrapper[4824]: E1209 09:47:46.898740 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:46Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:46 crc kubenswrapper[4824]: E1209 09:47:46.898893 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.900292 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.900313 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.900322 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.900336 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.900347 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:46Z","lastTransitionTime":"2025-12-09T09:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.910111 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.910173 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:47:46 crc kubenswrapper[4824]: I1209 09:47:46.910201 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:47:46 crc kubenswrapper[4824]: E1209 09:47:46.910249 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:47:46 crc kubenswrapper[4824]: E1209 09:47:46.910364 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:47:46 crc kubenswrapper[4824]: E1209 09:47:46.910434 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.005372 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.006272 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.006307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.006337 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.006353 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:47Z","lastTransitionTime":"2025-12-09T09:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.078182 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerStarted","Data":"db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c"} Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.078659 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerStarted","Data":"02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96"} Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.078677 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerStarted","Data":"d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8"} Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.078691 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerStarted","Data":"8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8"} Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.078703 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerStarted","Data":"64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f"} Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.078716 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerStarted","Data":"1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695"} Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.080062 4824 generic.go:334] "Generic (PLEG): container finished" podID="9e41bcf8-bacf-4eb1-a391-bbd86adedb32" containerID="660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b" exitCode=0 Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.080122 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" event={"ID":"9e41bcf8-bacf-4eb1-a391-bbd86adedb32","Type":"ContainerDied","Data":"660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b"} Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.082703 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerStarted","Data":"27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f"} Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.093217 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.109068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.109117 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.109131 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.109148 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.109160 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:47Z","lastTransitionTime":"2025-12-09T09:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.111938 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.125041 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.136515 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.152917 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.169350 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.184014 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.202261 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.212992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.213075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.213094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.213116 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.213130 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:47Z","lastTransitionTime":"2025-12-09T09:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.216495 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.227223 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.239696 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.267567 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.284052 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.298203 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.311991 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.316164 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.316197 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.316248 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.316273 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.316286 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:47Z","lastTransitionTime":"2025-12-09T09:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.324414 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.342476 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.364054 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.380300 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.399296 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.417826 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.418008 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-sxd45"] Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.418597 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-sxd45" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.420635 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.420915 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.420951 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.420964 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.420982 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.420996 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:47Z","lastTransitionTime":"2025-12-09T09:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.421023 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.421150 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.423038 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.435193 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.449483 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.462743 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.475135 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.486415 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.497653 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.510947 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.519149 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s87f\" (UniqueName: \"kubernetes.io/projected/40095fcc-84f8-4c61-8255-f76878bdc3bb-kube-api-access-4s87f\") pod \"node-ca-sxd45\" (UID: \"40095fcc-84f8-4c61-8255-f76878bdc3bb\") " pod="openshift-image-registry/node-ca-sxd45" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.519211 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/40095fcc-84f8-4c61-8255-f76878bdc3bb-serviceca\") pod \"node-ca-sxd45\" (UID: \"40095fcc-84f8-4c61-8255-f76878bdc3bb\") " pod="openshift-image-registry/node-ca-sxd45" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.519241 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/40095fcc-84f8-4c61-8255-f76878bdc3bb-host\") pod \"node-ca-sxd45\" (UID: \"40095fcc-84f8-4c61-8255-f76878bdc3bb\") " pod="openshift-image-registry/node-ca-sxd45" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.523610 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.524885 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.524921 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.524933 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.524953 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.524968 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:47Z","lastTransitionTime":"2025-12-09T09:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.534882 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.543653 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.561607 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.574142 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.589240 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.602362 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.616074 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.620606 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/40095fcc-84f8-4c61-8255-f76878bdc3bb-host\") pod \"node-ca-sxd45\" (UID: \"40095fcc-84f8-4c61-8255-f76878bdc3bb\") " pod="openshift-image-registry/node-ca-sxd45" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.620807 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4s87f\" (UniqueName: \"kubernetes.io/projected/40095fcc-84f8-4c61-8255-f76878bdc3bb-kube-api-access-4s87f\") pod \"node-ca-sxd45\" (UID: \"40095fcc-84f8-4c61-8255-f76878bdc3bb\") " pod="openshift-image-registry/node-ca-sxd45" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.620730 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/40095fcc-84f8-4c61-8255-f76878bdc3bb-host\") pod \"node-ca-sxd45\" (UID: \"40095fcc-84f8-4c61-8255-f76878bdc3bb\") " pod="openshift-image-registry/node-ca-sxd45" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.620926 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/40095fcc-84f8-4c61-8255-f76878bdc3bb-serviceca\") pod \"node-ca-sxd45\" (UID: \"40095fcc-84f8-4c61-8255-f76878bdc3bb\") " pod="openshift-image-registry/node-ca-sxd45" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.622279 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/40095fcc-84f8-4c61-8255-f76878bdc3bb-serviceca\") pod \"node-ca-sxd45\" (UID: \"40095fcc-84f8-4c61-8255-f76878bdc3bb\") " pod="openshift-image-registry/node-ca-sxd45" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.626902 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.626940 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.626952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.626970 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.627223 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:47Z","lastTransitionTime":"2025-12-09T09:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.640413 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.680094 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s87f\" (UniqueName: \"kubernetes.io/projected/40095fcc-84f8-4c61-8255-f76878bdc3bb-kube-api-access-4s87f\") pod \"node-ca-sxd45\" (UID: \"40095fcc-84f8-4c61-8255-f76878bdc3bb\") " pod="openshift-image-registry/node-ca-sxd45" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.695502 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.730086 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.730134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.730143 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.730160 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.730180 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:47Z","lastTransitionTime":"2025-12-09T09:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.733186 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-sxd45" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.744915 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: W1209 09:47:47.755615 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40095fcc_84f8_4c61_8255_f76878bdc3bb.slice/crio-d915f04e238dce4e65171315687f88b0b2a164d49a55551a6b551a81623975b7 WatchSource:0}: Error finding container d915f04e238dce4e65171315687f88b0b2a164d49a55551a6b551a81623975b7: Status 404 returned error can't find the container with id d915f04e238dce4e65171315687f88b0b2a164d49a55551a6b551a81623975b7 Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.778644 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.817340 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.833469 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.833504 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.833513 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.833528 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.833537 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:47Z","lastTransitionTime":"2025-12-09T09:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.850578 4824 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.878461 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.915473 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:47Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.935815 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.935854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.935865 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.935879 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:47 crc kubenswrapper[4824]: I1209 09:47:47.935888 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:47Z","lastTransitionTime":"2025-12-09T09:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.039539 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.039597 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.039611 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.039633 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.039647 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:48Z","lastTransitionTime":"2025-12-09T09:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.088711 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-sxd45" event={"ID":"40095fcc-84f8-4c61-8255-f76878bdc3bb","Type":"ContainerStarted","Data":"43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49"} Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.088850 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-sxd45" event={"ID":"40095fcc-84f8-4c61-8255-f76878bdc3bb","Type":"ContainerStarted","Data":"d915f04e238dce4e65171315687f88b0b2a164d49a55551a6b551a81623975b7"} Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.091599 4824 generic.go:334] "Generic (PLEG): container finished" podID="9e41bcf8-bacf-4eb1-a391-bbd86adedb32" containerID="579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f" exitCode=0 Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.091666 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" event={"ID":"9e41bcf8-bacf-4eb1-a391-bbd86adedb32","Type":"ContainerDied","Data":"579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f"} Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.102419 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.121660 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.143760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.143815 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.143825 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.143840 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.143849 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:48Z","lastTransitionTime":"2025-12-09T09:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.146997 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.158225 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.169905 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.182301 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.201999 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.238118 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.254144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.254188 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.254202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.254222 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.254236 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:48Z","lastTransitionTime":"2025-12-09T09:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.280330 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.318989 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.357260 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.357305 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.357316 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.357332 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.357343 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:48Z","lastTransitionTime":"2025-12-09T09:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.359868 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.396906 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.438028 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.459276 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.459322 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.459333 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.459355 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.459368 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:48Z","lastTransitionTime":"2025-12-09T09:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.477478 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.518498 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.558888 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.561863 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.561895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.561906 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.561924 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.561937 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:48Z","lastTransitionTime":"2025-12-09T09:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.597896 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.634394 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.664507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.664551 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.664561 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.664575 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.664585 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:48Z","lastTransitionTime":"2025-12-09T09:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.675367 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.716584 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.759153 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.767290 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.767349 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.767366 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.767387 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.767404 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:48Z","lastTransitionTime":"2025-12-09T09:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.800091 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.842377 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.870039 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.870073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.870082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.870099 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.870117 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:48Z","lastTransitionTime":"2025-12-09T09:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.881983 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.910232 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.910299 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.910246 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:47:48 crc kubenswrapper[4824]: E1209 09:47:48.910437 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:47:48 crc kubenswrapper[4824]: E1209 09:47:48.910584 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:47:48 crc kubenswrapper[4824]: E1209 09:47:48.910880 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.919633 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.957239 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:48Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.972773 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.972858 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.972870 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.972888 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:48 crc kubenswrapper[4824]: I1209 09:47:48.972899 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:48Z","lastTransitionTime":"2025-12-09T09:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.007157 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:49Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.040103 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:49Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.074921 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.074958 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.074967 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.074980 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.074989 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:49Z","lastTransitionTime":"2025-12-09T09:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.079829 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:49Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.098427 4824 generic.go:334] "Generic (PLEG): container finished" podID="9e41bcf8-bacf-4eb1-a391-bbd86adedb32" containerID="ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f" exitCode=0 Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.098489 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" event={"ID":"9e41bcf8-bacf-4eb1-a391-bbd86adedb32","Type":"ContainerDied","Data":"ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f"} Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.120447 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:49Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.156100 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:49Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.177578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.177621 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.177630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.177648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.177658 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:49Z","lastTransitionTime":"2025-12-09T09:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.203853 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:49Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.240598 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:49Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.277183 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:49Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.279428 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.279458 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.279467 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.279483 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.279500 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:49Z","lastTransitionTime":"2025-12-09T09:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.317619 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:49Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.356929 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:49Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.382110 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.382146 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.382159 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.382176 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.382189 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:49Z","lastTransitionTime":"2025-12-09T09:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.397715 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:49Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.436706 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:49Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.474731 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:49Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.484538 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.484589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.484605 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.484628 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.484648 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:49Z","lastTransitionTime":"2025-12-09T09:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.517339 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:49Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.559552 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:49Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.588400 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.588457 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.588468 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.588488 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.588503 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:49Z","lastTransitionTime":"2025-12-09T09:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.600572 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:49Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.639712 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:49Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.682678 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:49Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.691513 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.691577 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.691590 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.691612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.691841 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:49Z","lastTransitionTime":"2025-12-09T09:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.743169 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:49Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.794285 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.794339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.794351 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.794369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.794386 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:49Z","lastTransitionTime":"2025-12-09T09:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.897957 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.898016 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.898028 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.898046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:49 crc kubenswrapper[4824]: I1209 09:47:49.898057 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:49Z","lastTransitionTime":"2025-12-09T09:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.000956 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.001010 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.001022 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.001043 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.001056 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:50Z","lastTransitionTime":"2025-12-09T09:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.103085 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.103134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.103147 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.103167 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.103180 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:50Z","lastTransitionTime":"2025-12-09T09:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.106839 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerStarted","Data":"c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f"} Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.109237 4824 generic.go:334] "Generic (PLEG): container finished" podID="9e41bcf8-bacf-4eb1-a391-bbd86adedb32" containerID="45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d" exitCode=0 Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.109271 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" event={"ID":"9e41bcf8-bacf-4eb1-a391-bbd86adedb32","Type":"ContainerDied","Data":"45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d"} Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.125236 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:50Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.142966 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:50Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.164379 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:50Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.181380 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:50Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.196348 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:50Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.204956 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.204996 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.205006 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.205023 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.205034 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:50Z","lastTransitionTime":"2025-12-09T09:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.212556 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:50Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.231776 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:50Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.254179 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:50Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.275192 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:50Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.293764 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:50Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.308347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.308398 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.308414 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.308439 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.308457 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:50Z","lastTransitionTime":"2025-12-09T09:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.309892 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:50Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.322078 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:50Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.337166 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:50Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.355073 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:50Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.369546 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:50Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.412069 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.412113 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.412125 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.412143 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.412157 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:50Z","lastTransitionTime":"2025-12-09T09:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.514550 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.514598 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.514616 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.514639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.514651 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:50Z","lastTransitionTime":"2025-12-09T09:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.550203 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:47:50 crc kubenswrapper[4824]: E1209 09:47:50.550440 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:47:58.550403633 +0000 UTC m=+34.884908300 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.617466 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.617508 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.617520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.617535 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.617545 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:50Z","lastTransitionTime":"2025-12-09T09:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.651297 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.651357 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.651377 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.651407 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:47:50 crc kubenswrapper[4824]: E1209 09:47:50.651516 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 09:47:50 crc kubenswrapper[4824]: E1209 09:47:50.651558 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 09:47:50 crc kubenswrapper[4824]: E1209 09:47:50.651572 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 09:47:50 crc kubenswrapper[4824]: E1209 09:47:50.651596 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 09:47:50 crc kubenswrapper[4824]: E1209 09:47:50.651612 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:47:50 crc kubenswrapper[4824]: E1209 09:47:50.651622 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 09:47:58.651596207 +0000 UTC m=+34.986100874 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 09:47:50 crc kubenswrapper[4824]: E1209 09:47:50.651646 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 09:47:58.651633709 +0000 UTC m=+34.986138386 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 09:47:50 crc kubenswrapper[4824]: E1209 09:47:50.651661 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 09:47:58.651654439 +0000 UTC m=+34.986159106 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:47:50 crc kubenswrapper[4824]: E1209 09:47:50.651667 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 09:47:50 crc kubenswrapper[4824]: E1209 09:47:50.651707 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 09:47:50 crc kubenswrapper[4824]: E1209 09:47:50.651734 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:47:50 crc kubenswrapper[4824]: E1209 09:47:50.651828 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 09:47:58.651803424 +0000 UTC m=+34.986308111 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.720184 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.720577 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.720662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.720748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.720887 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:50Z","lastTransitionTime":"2025-12-09T09:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.823216 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.823283 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.823297 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.823313 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.823324 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:50Z","lastTransitionTime":"2025-12-09T09:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.910398 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.910546 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:47:50 crc kubenswrapper[4824]: E1209 09:47:50.910695 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.910760 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:47:50 crc kubenswrapper[4824]: E1209 09:47:50.910852 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:47:50 crc kubenswrapper[4824]: E1209 09:47:50.910966 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.926310 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.926338 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.926347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.926362 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:50 crc kubenswrapper[4824]: I1209 09:47:50.926372 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:50Z","lastTransitionTime":"2025-12-09T09:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.028594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.028653 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.028665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.028687 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.028701 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:51Z","lastTransitionTime":"2025-12-09T09:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.116301 4824 generic.go:334] "Generic (PLEG): container finished" podID="9e41bcf8-bacf-4eb1-a391-bbd86adedb32" containerID="62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe" exitCode=0 Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.116356 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" event={"ID":"9e41bcf8-bacf-4eb1-a391-bbd86adedb32","Type":"ContainerDied","Data":"62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe"} Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.131669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.131724 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.131735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.131755 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.131766 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:51Z","lastTransitionTime":"2025-12-09T09:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.133551 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:51Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.159454 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:51Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.180522 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:51Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.195977 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:51Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.210645 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:51Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.225850 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:51Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.234402 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.234448 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.234458 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.234477 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.234488 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:51Z","lastTransitionTime":"2025-12-09T09:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.245900 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:51Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.260446 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:51Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.274255 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:51Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.288911 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:51Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.301466 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:51Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.312987 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:51Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.325911 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:51Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.337509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.337549 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.337558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.337576 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.337585 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:51Z","lastTransitionTime":"2025-12-09T09:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.338651 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:51Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.359830 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:51Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.439937 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.439984 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.439996 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.440014 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.440028 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:51Z","lastTransitionTime":"2025-12-09T09:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.546581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.546634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.546646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.546667 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.546677 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:51Z","lastTransitionTime":"2025-12-09T09:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.649841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.649910 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.649927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.649952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.649963 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:51Z","lastTransitionTime":"2025-12-09T09:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.753076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.753133 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.753144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.753161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.753174 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:51Z","lastTransitionTime":"2025-12-09T09:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.856402 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.856440 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.856451 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.856469 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.856481 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:51Z","lastTransitionTime":"2025-12-09T09:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.959110 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.959169 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.959182 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.959202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:51 crc kubenswrapper[4824]: I1209 09:47:51.959217 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:51Z","lastTransitionTime":"2025-12-09T09:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.062380 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.062432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.062441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.062456 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.062467 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:52Z","lastTransitionTime":"2025-12-09T09:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.126099 4824 generic.go:334] "Generic (PLEG): container finished" podID="9e41bcf8-bacf-4eb1-a391-bbd86adedb32" containerID="6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866" exitCode=0 Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.126162 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" event={"ID":"9e41bcf8-bacf-4eb1-a391-bbd86adedb32","Type":"ContainerDied","Data":"6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866"} Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.133433 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerStarted","Data":"16621af89dcbe1805241c84c225b67023c3ec7d9f8a42c31ed7bf7a01fc1d9ee"} Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.133838 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.140888 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.159591 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.160289 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.164741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.164768 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.164789 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.164802 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.164811 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:52Z","lastTransitionTime":"2025-12-09T09:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.181080 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.195561 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.208869 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.221642 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.235157 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.255917 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.267292 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.267328 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.267338 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.267352 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.267363 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:52Z","lastTransitionTime":"2025-12-09T09:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.270300 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.283894 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.297362 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.307195 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.319513 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.332854 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.345252 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.360496 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.370994 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.371045 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.371057 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.371079 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.371093 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:52Z","lastTransitionTime":"2025-12-09T09:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.376505 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.396508 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16621af89dcbe1805241c84c225b67023c3ec7d9f8a42c31ed7bf7a01fc1d9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.411183 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.430147 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.440830 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.452152 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.473164 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.474872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.474920 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.474930 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.474948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.474960 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:52Z","lastTransitionTime":"2025-12-09T09:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.492459 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.510117 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.527906 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.538222 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.550950 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.566005 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.577315 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.577370 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.577385 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.577406 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.577423 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:52Z","lastTransitionTime":"2025-12-09T09:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.579103 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:52Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.679916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.679971 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.679985 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.680005 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.680022 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:52Z","lastTransitionTime":"2025-12-09T09:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.782310 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.782370 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.782393 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.782415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.782430 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:52Z","lastTransitionTime":"2025-12-09T09:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.884898 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.884925 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.884934 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.884947 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.884956 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:52Z","lastTransitionTime":"2025-12-09T09:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.909851 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.909852 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.910097 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:47:52 crc kubenswrapper[4824]: E1209 09:47:52.910254 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:47:52 crc kubenswrapper[4824]: E1209 09:47:52.910431 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:47:52 crc kubenswrapper[4824]: E1209 09:47:52.910556 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.987492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.987573 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.987593 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.987620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:52 crc kubenswrapper[4824]: I1209 09:47:52.987642 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:52Z","lastTransitionTime":"2025-12-09T09:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.090592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.090663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.090682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.090705 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.090723 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:53Z","lastTransitionTime":"2025-12-09T09:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.140967 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" event={"ID":"9e41bcf8-bacf-4eb1-a391-bbd86adedb32","Type":"ContainerStarted","Data":"161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181"} Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.141052 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.141706 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.158909 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.166669 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.173444 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.188699 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.193581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.193638 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.193650 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.193671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.193687 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:53Z","lastTransitionTime":"2025-12-09T09:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.204067 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.222166 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.242098 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16621af89dcbe1805241c84c225b67023c3ec7d9f8a42c31ed7bf7a01fc1d9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.256836 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.272300 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.297024 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.297101 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.297115 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.297139 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.297153 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:53Z","lastTransitionTime":"2025-12-09T09:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.297223 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.319443 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.334264 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.348511 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.362105 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.384203 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.395829 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.399546 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.399576 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.399586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.399603 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.399615 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:53Z","lastTransitionTime":"2025-12-09T09:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.425010 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.443156 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.470199 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.488862 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.498455 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.505424 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.505468 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.505482 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.505500 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.505512 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:53Z","lastTransitionTime":"2025-12-09T09:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.513599 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.526569 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.550450 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.568710 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.588532 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.625995 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.627313 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.627427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.627541 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.627631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.627722 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:53Z","lastTransitionTime":"2025-12-09T09:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.641350 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.656662 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.671882 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.694655 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16621af89dcbe1805241c84c225b67023c3ec7d9f8a42c31ed7bf7a01fc1d9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.730144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.730210 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.730227 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.730253 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.730270 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:53Z","lastTransitionTime":"2025-12-09T09:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.833345 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.833445 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.833457 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.833477 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.833489 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:53Z","lastTransitionTime":"2025-12-09T09:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.927410 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.935640 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.935680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.935691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.935710 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.935726 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:53Z","lastTransitionTime":"2025-12-09T09:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.944231 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.961072 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16621af89dcbe1805241c84c225b67023c3ec7d9f8a42c31ed7bf7a01fc1d9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.980308 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:53 crc kubenswrapper[4824]: I1209 09:47:53.994455 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:53Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.006268 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:54Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.019607 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:54Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.034074 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:54Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.038265 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.038302 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.038316 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.038336 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.038350 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:54Z","lastTransitionTime":"2025-12-09T09:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.051153 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:54Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.064266 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:54Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.078589 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:54Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.097346 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:54Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.110050 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:54Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.121394 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:54Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.137937 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:54Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.141068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.141102 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.141114 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.141132 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.141144 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:54Z","lastTransitionTime":"2025-12-09T09:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.143392 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.244261 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.244291 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.244301 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.244315 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.244323 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:54Z","lastTransitionTime":"2025-12-09T09:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.347817 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.347872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.347888 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.347911 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.347925 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:54Z","lastTransitionTime":"2025-12-09T09:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.451342 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.451401 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.451415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.451433 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.451447 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:54Z","lastTransitionTime":"2025-12-09T09:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.476629 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.554666 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.555041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.555054 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.555076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.555090 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:54Z","lastTransitionTime":"2025-12-09T09:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.658160 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.658199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.658214 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.658233 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.658245 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:54Z","lastTransitionTime":"2025-12-09T09:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.761151 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.761191 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.761200 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.761216 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.761225 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:54Z","lastTransitionTime":"2025-12-09T09:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.863198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.863235 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.863247 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.863263 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.863276 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:54Z","lastTransitionTime":"2025-12-09T09:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.909544 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.909595 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.909561 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:47:54 crc kubenswrapper[4824]: E1209 09:47:54.909718 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:47:54 crc kubenswrapper[4824]: E1209 09:47:54.909791 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:47:54 crc kubenswrapper[4824]: E1209 09:47:54.909865 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.966072 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.966115 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.966131 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.966152 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:54 crc kubenswrapper[4824]: I1209 09:47:54.966168 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:54Z","lastTransitionTime":"2025-12-09T09:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.068982 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.069076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.069091 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.069112 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.069132 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:55Z","lastTransitionTime":"2025-12-09T09:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.150110 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vvxck_13620d58-6925-48c7-bbc7-f34e92922480/ovnkube-controller/0.log" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.153823 4824 generic.go:334] "Generic (PLEG): container finished" podID="13620d58-6925-48c7-bbc7-f34e92922480" containerID="16621af89dcbe1805241c84c225b67023c3ec7d9f8a42c31ed7bf7a01fc1d9ee" exitCode=1 Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.153872 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerDied","Data":"16621af89dcbe1805241c84c225b67023c3ec7d9f8a42c31ed7bf7a01fc1d9ee"} Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.155344 4824 scope.go:117] "RemoveContainer" containerID="16621af89dcbe1805241c84c225b67023c3ec7d9f8a42c31ed7bf7a01fc1d9ee" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.176972 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.177042 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.177058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.177084 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.177100 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:55Z","lastTransitionTime":"2025-12-09T09:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.182147 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:55Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.202804 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16621af89dcbe1805241c84c225b67023c3ec7d9f8a42c31ed7bf7a01fc1d9ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16621af89dcbe1805241c84c225b67023c3ec7d9f8a42c31ed7bf7a01fc1d9ee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:47:54Z\\\",\\\"message\\\":\\\"ressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 09:47:54.235037 6115 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1209 09:47:54.235074 6115 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1209 09:47:54.235088 6115 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1209 09:47:54.235092 6115 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1209 09:47:54.235116 6115 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1209 09:47:54.235190 6115 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1209 09:47:54.235197 6115 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1209 09:47:54.235237 6115 factory.go:656] Stopping watch factory\\\\nI1209 09:47:54.235238 6115 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1209 09:47:54.235248 6115 ovnkube.go:599] Stopped ovnkube\\\\nI1209 09:47:54.235254 6115 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1209 09:47:54.235263 6115 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1209 09:47:54.235272 6115 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1209 09:47:54.235140 6115 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1209 09:47:54.235154 6115 handler.go:208] Removed *v1.Namespace event handler \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:55Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.215645 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:55Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.238211 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:55Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.253820 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:55Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.269340 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:55Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.279672 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.279709 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.279718 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.279733 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.279743 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:55Z","lastTransitionTime":"2025-12-09T09:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.281205 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:55Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.292768 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:55Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.305148 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:55Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.317717 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:55Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.332105 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:55Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.349379 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:55Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.365649 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:55Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.378489 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:55Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.384196 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.384261 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.384273 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.384291 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.384631 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:55Z","lastTransitionTime":"2025-12-09T09:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.395124 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:55Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.487547 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.487589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.487600 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.487615 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.487626 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:55Z","lastTransitionTime":"2025-12-09T09:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.590549 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.590605 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.590637 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.590663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.590682 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:55Z","lastTransitionTime":"2025-12-09T09:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.693874 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.693931 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.693941 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.693961 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.693971 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:55Z","lastTransitionTime":"2025-12-09T09:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.796484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.796563 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.796573 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.796596 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.796607 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:55Z","lastTransitionTime":"2025-12-09T09:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.903159 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.903211 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.903228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.903247 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.903260 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:55Z","lastTransitionTime":"2025-12-09T09:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:55 crc kubenswrapper[4824]: I1209 09:47:55.996860 4824 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.006427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.006480 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.006491 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.006509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.006519 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:56Z","lastTransitionTime":"2025-12-09T09:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.109087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.109141 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.109154 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.109170 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.109181 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:56Z","lastTransitionTime":"2025-12-09T09:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.159078 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vvxck_13620d58-6925-48c7-bbc7-f34e92922480/ovnkube-controller/0.log" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.161515 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerStarted","Data":"102e8ec6dccd87b7bf834dfa6d7c2d18f2e5b9bfe7921d3ecc060056352e3b27"} Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.162260 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.190265 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:56Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.204415 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:56Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.211814 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.211864 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.211876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.211893 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.211904 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:56Z","lastTransitionTime":"2025-12-09T09:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.221348 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:56Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.233916 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:56Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.246826 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:56Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.262108 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:56Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.276949 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:56Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.291402 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:56Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.302977 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:56Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.314561 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.314596 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.314606 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.314620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.314630 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:56Z","lastTransitionTime":"2025-12-09T09:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.318535 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:56Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.331945 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:56Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.345496 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:56Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.357463 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:56Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.371925 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:56Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.397934 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://102e8ec6dccd87b7bf834dfa6d7c2d18f2e5b9bfe7921d3ecc060056352e3b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16621af89dcbe1805241c84c225b67023c3ec7d9f8a42c31ed7bf7a01fc1d9ee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:47:54Z\\\",\\\"message\\\":\\\"ressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 09:47:54.235037 6115 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1209 09:47:54.235074 6115 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1209 09:47:54.235088 6115 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1209 09:47:54.235092 6115 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1209 09:47:54.235116 6115 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1209 09:47:54.235190 6115 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1209 09:47:54.235197 6115 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1209 09:47:54.235237 6115 factory.go:656] Stopping watch factory\\\\nI1209 09:47:54.235238 6115 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1209 09:47:54.235248 6115 ovnkube.go:599] Stopped ovnkube\\\\nI1209 09:47:54.235254 6115 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1209 09:47:54.235263 6115 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1209 09:47:54.235272 6115 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1209 09:47:54.235140 6115 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1209 09:47:54.235154 6115 handler.go:208] Removed *v1.Namespace event handler \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:56Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.418061 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.418383 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.418465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.418624 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.418720 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:56Z","lastTransitionTime":"2025-12-09T09:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.523021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.523092 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.523111 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.523138 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.523162 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:56Z","lastTransitionTime":"2025-12-09T09:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.625854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.626305 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.626407 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.626512 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.626612 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:56Z","lastTransitionTime":"2025-12-09T09:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.729893 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.730331 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.730427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.730526 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.730613 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:56Z","lastTransitionTime":"2025-12-09T09:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.833407 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.833649 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.833715 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.833845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.833942 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:56Z","lastTransitionTime":"2025-12-09T09:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.910016 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.910156 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.910534 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:47:56 crc kubenswrapper[4824]: E1209 09:47:56.910760 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:47:56 crc kubenswrapper[4824]: E1209 09:47:56.910951 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:47:56 crc kubenswrapper[4824]: E1209 09:47:56.911036 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.938021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.938082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.938103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.938132 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:56 crc kubenswrapper[4824]: I1209 09:47:56.938157 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:56Z","lastTransitionTime":"2025-12-09T09:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.043381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.043432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.043469 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.043493 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.043509 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:57Z","lastTransitionTime":"2025-12-09T09:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.148184 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.148249 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.148263 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.148288 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.148308 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:57Z","lastTransitionTime":"2025-12-09T09:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.166827 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vvxck_13620d58-6925-48c7-bbc7-f34e92922480/ovnkube-controller/1.log" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.167501 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vvxck_13620d58-6925-48c7-bbc7-f34e92922480/ovnkube-controller/0.log" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.170004 4824 generic.go:334] "Generic (PLEG): container finished" podID="13620d58-6925-48c7-bbc7-f34e92922480" containerID="102e8ec6dccd87b7bf834dfa6d7c2d18f2e5b9bfe7921d3ecc060056352e3b27" exitCode=1 Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.170050 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerDied","Data":"102e8ec6dccd87b7bf834dfa6d7c2d18f2e5b9bfe7921d3ecc060056352e3b27"} Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.170115 4824 scope.go:117] "RemoveContainer" containerID="16621af89dcbe1805241c84c225b67023c3ec7d9f8a42c31ed7bf7a01fc1d9ee" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.171653 4824 scope.go:117] "RemoveContainer" containerID="102e8ec6dccd87b7bf834dfa6d7c2d18f2e5b9bfe7921d3ecc060056352e3b27" Dec 09 09:47:57 crc kubenswrapper[4824]: E1209 09:47:57.171902 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-vvxck_openshift-ovn-kubernetes(13620d58-6925-48c7-bbc7-f34e92922480)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" podUID="13620d58-6925-48c7-bbc7-f34e92922480" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.187917 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.200726 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.227294 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.247020 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.251252 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.251302 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.251313 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.251333 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.251348 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:57Z","lastTransitionTime":"2025-12-09T09:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.265173 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.303989 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.304061 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.304081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.304115 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.304146 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:57Z","lastTransitionTime":"2025-12-09T09:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.316812 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: E1209 09:47:57.318045 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.321282 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.321342 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.321357 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.321402 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.321414 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:57Z","lastTransitionTime":"2025-12-09T09:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.330090 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: E1209 09:47:57.333961 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.338535 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.338591 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.338609 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.338643 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.338655 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:57Z","lastTransitionTime":"2025-12-09T09:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.342634 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.353163 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: E1209 09:47:57.353223 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.357231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.357266 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.357282 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.357302 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.357315 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:57Z","lastTransitionTime":"2025-12-09T09:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.369320 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: E1209 09:47:57.369824 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.374750 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.374815 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.374827 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.374867 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.374881 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:57Z","lastTransitionTime":"2025-12-09T09:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.384909 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: E1209 09:47:57.395918 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: E1209 09:47:57.396118 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.396357 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx"] Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.397146 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.398889 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.398939 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.398954 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.398971 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.398986 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:57Z","lastTransitionTime":"2025-12-09T09:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.400990 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.401406 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.401998 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.414940 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.430480 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.449662 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://102e8ec6dccd87b7bf834dfa6d7c2d18f2e5b9bfe7921d3ecc060056352e3b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16621af89dcbe1805241c84c225b67023c3ec7d9f8a42c31ed7bf7a01fc1d9ee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:47:54Z\\\",\\\"message\\\":\\\"ressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 09:47:54.235037 6115 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1209 09:47:54.235074 6115 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1209 09:47:54.235088 6115 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1209 09:47:54.235092 6115 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1209 09:47:54.235116 6115 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1209 09:47:54.235190 6115 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1209 09:47:54.235197 6115 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1209 09:47:54.235237 6115 factory.go:656] Stopping watch factory\\\\nI1209 09:47:54.235238 6115 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1209 09:47:54.235248 6115 ovnkube.go:599] Stopped ovnkube\\\\nI1209 09:47:54.235254 6115 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1209 09:47:54.235263 6115 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1209 09:47:54.235272 6115 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1209 09:47:54.235140 6115 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1209 09:47:54.235154 6115 handler.go:208] Removed *v1.Namespace event handler \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://102e8ec6dccd87b7bf834dfa6d7c2d18f2e5b9bfe7921d3ecc060056352e3b27\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:47:56Z\\\",\\\"message\\\":\\\"false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ba175bbe-5cc4-47e6-a32d-57693e1320bd}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 09:47:56.026026 6254 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager/kube-controller-manager]} name:Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ba175bbe-5cc4-47e6-a32d-57693e1320bd}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 09:47:56.026256 6254 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1209 09:47:56.026337 6254 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1209 09:47:56.026374 6254 ovnkube.go:599] Stopped ovnkube\\\\nI1209 09:47:56.026400 6254 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1209 09:47:56.026471 6254 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.464210 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.479925 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.498244 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.502800 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.502847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.502860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.502881 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.502898 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:57Z","lastTransitionTime":"2025-12-09T09:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.520152 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://102e8ec6dccd87b7bf834dfa6d7c2d18f2e5b9bfe7921d3ecc060056352e3b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16621af89dcbe1805241c84c225b67023c3ec7d9f8a42c31ed7bf7a01fc1d9ee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:47:54Z\\\",\\\"message\\\":\\\"ressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 09:47:54.235037 6115 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1209 09:47:54.235074 6115 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1209 09:47:54.235088 6115 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1209 09:47:54.235092 6115 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1209 09:47:54.235116 6115 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1209 09:47:54.235190 6115 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1209 09:47:54.235197 6115 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1209 09:47:54.235237 6115 factory.go:656] Stopping watch factory\\\\nI1209 09:47:54.235238 6115 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1209 09:47:54.235248 6115 ovnkube.go:599] Stopped ovnkube\\\\nI1209 09:47:54.235254 6115 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1209 09:47:54.235263 6115 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1209 09:47:54.235272 6115 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1209 09:47:54.235140 6115 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1209 09:47:54.235154 6115 handler.go:208] Removed *v1.Namespace event handler \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://102e8ec6dccd87b7bf834dfa6d7c2d18f2e5b9bfe7921d3ecc060056352e3b27\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:47:56Z\\\",\\\"message\\\":\\\"false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ba175bbe-5cc4-47e6-a32d-57693e1320bd}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 09:47:56.026026 6254 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager/kube-controller-manager]} name:Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ba175bbe-5cc4-47e6-a32d-57693e1320bd}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 09:47:56.026256 6254 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1209 09:47:56.026337 6254 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1209 09:47:56.026374 6254 ovnkube.go:599] Stopped ovnkube\\\\nI1209 09:47:56.026400 6254 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1209 09:47:56.026471 6254 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.522678 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5-env-overrides\") pod \"ovnkube-control-plane-749d76644c-gcrqx\" (UID: \"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.522735 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-gcrqx\" (UID: \"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.522848 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-256nl\" (UniqueName: \"kubernetes.io/projected/1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5-kube-api-access-256nl\") pod \"ovnkube-control-plane-749d76644c-gcrqx\" (UID: \"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.522929 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-gcrqx\" (UID: \"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.535041 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gcrqx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.549828 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.569762 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.586014 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.600495 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.605494 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.605544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.605565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.605599 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.605616 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:57Z","lastTransitionTime":"2025-12-09T09:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.614519 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.624326 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5-env-overrides\") pod \"ovnkube-control-plane-749d76644c-gcrqx\" (UID: \"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.624363 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-gcrqx\" (UID: \"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.624438 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-256nl\" (UniqueName: \"kubernetes.io/projected/1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5-kube-api-access-256nl\") pod \"ovnkube-control-plane-749d76644c-gcrqx\" (UID: \"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.624470 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-gcrqx\" (UID: \"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.625319 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5-env-overrides\") pod \"ovnkube-control-plane-749d76644c-gcrqx\" (UID: \"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.625402 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-gcrqx\" (UID: \"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.627844 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.632534 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-gcrqx\" (UID: \"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.641335 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-256nl\" (UniqueName: \"kubernetes.io/projected/1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5-kube-api-access-256nl\") pod \"ovnkube-control-plane-749d76644c-gcrqx\" (UID: \"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.641732 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.662375 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.674268 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.690380 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.703273 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:57Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.708217 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.708261 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.708271 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.708288 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.708297 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:57Z","lastTransitionTime":"2025-12-09T09:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.717559 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.811583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.811619 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.811629 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.811646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.811656 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:57Z","lastTransitionTime":"2025-12-09T09:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.914040 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.914108 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.914125 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.914146 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:57 crc kubenswrapper[4824]: I1209 09:47:57.914162 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:57Z","lastTransitionTime":"2025-12-09T09:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.017959 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.018006 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.018018 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.018034 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.018094 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:58Z","lastTransitionTime":"2025-12-09T09:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.122024 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.122119 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.122130 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.122185 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.122199 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:58Z","lastTransitionTime":"2025-12-09T09:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.177224 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vvxck_13620d58-6925-48c7-bbc7-f34e92922480/ovnkube-controller/1.log" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.183758 4824 scope.go:117] "RemoveContainer" containerID="102e8ec6dccd87b7bf834dfa6d7c2d18f2e5b9bfe7921d3ecc060056352e3b27" Dec 09 09:47:58 crc kubenswrapper[4824]: E1209 09:47:58.183979 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-vvxck_openshift-ovn-kubernetes(13620d58-6925-48c7-bbc7-f34e92922480)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" podUID="13620d58-6925-48c7-bbc7-f34e92922480" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.186682 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" event={"ID":"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5","Type":"ContainerStarted","Data":"daad2ae578b4f337ceb3298b012b0262903cebe4e9e9584513cf2f31fbcec3c9"} Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.186735 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" event={"ID":"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5","Type":"ContainerStarted","Data":"f8d3936e16db7382c8bc6153743d9b29037e13c68aa3a2cf02b21d1ea9dd9ff1"} Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.201253 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.216435 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.225143 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.225181 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.225193 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.225213 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.225227 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:58Z","lastTransitionTime":"2025-12-09T09:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.233269 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.248033 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.264125 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.288941 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://102e8ec6dccd87b7bf834dfa6d7c2d18f2e5b9bfe7921d3ecc060056352e3b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://102e8ec6dccd87b7bf834dfa6d7c2d18f2e5b9bfe7921d3ecc060056352e3b27\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:47:56Z\\\",\\\"message\\\":\\\"false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ba175bbe-5cc4-47e6-a32d-57693e1320bd}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 09:47:56.026026 6254 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager/kube-controller-manager]} name:Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ba175bbe-5cc4-47e6-a32d-57693e1320bd}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 09:47:56.026256 6254 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1209 09:47:56.026337 6254 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1209 09:47:56.026374 6254 ovnkube.go:599] Stopped ovnkube\\\\nI1209 09:47:56.026400 6254 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1209 09:47:56.026471 6254 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-vvxck_openshift-ovn-kubernetes(13620d58-6925-48c7-bbc7-f34e92922480)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.301427 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gcrqx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.320952 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.327870 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.327926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.327939 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.327958 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.327980 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:58Z","lastTransitionTime":"2025-12-09T09:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.339262 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.356196 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.371532 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.394619 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.412548 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.425944 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.429941 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.429980 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.429996 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.430015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.430028 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:58Z","lastTransitionTime":"2025-12-09T09:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.440095 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.452574 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.499219 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-5tn52"] Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.500008 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:47:58 crc kubenswrapper[4824]: E1209 09:47:58.500113 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.517370 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.532618 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.532670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.532679 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.532698 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.532708 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:58Z","lastTransitionTime":"2025-12-09T09:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.539477 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://102e8ec6dccd87b7bf834dfa6d7c2d18f2e5b9bfe7921d3ecc060056352e3b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://102e8ec6dccd87b7bf834dfa6d7c2d18f2e5b9bfe7921d3ecc060056352e3b27\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:47:56Z\\\",\\\"message\\\":\\\"false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ba175bbe-5cc4-47e6-a32d-57693e1320bd}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 09:47:56.026026 6254 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager/kube-controller-manager]} name:Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ba175bbe-5cc4-47e6-a32d-57693e1320bd}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 09:47:56.026256 6254 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1209 09:47:56.026337 6254 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1209 09:47:56.026374 6254 ovnkube.go:599] Stopped ovnkube\\\\nI1209 09:47:56.026400 6254 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1209 09:47:56.026471 6254 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-vvxck_openshift-ovn-kubernetes(13620d58-6925-48c7-bbc7-f34e92922480)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.561067 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gcrqx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.574512 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5tn52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5tn52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.585769 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.604490 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.617581 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.632732 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.634832 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.634859 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.634868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.634882 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.634894 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:58Z","lastTransitionTime":"2025-12-09T09:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.636134 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:47:58 crc kubenswrapper[4824]: E1209 09:47:58.636307 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:48:14.636282276 +0000 UTC m=+50.970786983 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.636437 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psr7g\" (UniqueName: \"kubernetes.io/projected/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-kube-api-access-psr7g\") pod \"network-metrics-daemon-5tn52\" (UID: \"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\") " pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.636519 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs\") pod \"network-metrics-daemon-5tn52\" (UID: \"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\") " pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.648479 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.665137 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.684361 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.698099 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.708105 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.720905 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.737977 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.738021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.738031 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.738033 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psr7g\" (UniqueName: \"kubernetes.io/projected/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-kube-api-access-psr7g\") pod \"network-metrics-daemon-5tn52\" (UID: \"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\") " pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.738083 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.738117 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs\") pod \"network-metrics-daemon-5tn52\" (UID: \"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\") " pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.738135 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.738153 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.738175 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:47:58 crc kubenswrapper[4824]: E1209 09:47:58.738294 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 09:47:58 crc kubenswrapper[4824]: E1209 09:47:58.738345 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 09:48:14.738329136 +0000 UTC m=+51.072833803 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 09:47:58 crc kubenswrapper[4824]: E1209 09:47:58.738476 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 09:47:58 crc kubenswrapper[4824]: E1209 09:47:58.738501 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 09:47:58 crc kubenswrapper[4824]: E1209 09:47:58.738517 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:47:58 crc kubenswrapper[4824]: E1209 09:47:58.738576 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 09:48:14.738552273 +0000 UTC m=+51.073057110 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:47:58 crc kubenswrapper[4824]: E1209 09:47:58.738671 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 09:47:58 crc kubenswrapper[4824]: E1209 09:47:58.738703 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs podName:fcbce9ea-3d55-4e09-b5bc-d5bb8c472219 nodeName:}" failed. No retries permitted until 2025-12-09 09:47:59.238693468 +0000 UTC m=+35.573198135 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs") pod "network-metrics-daemon-5tn52" (UID: "fcbce9ea-3d55-4e09-b5bc-d5bb8c472219") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 09:47:58 crc kubenswrapper[4824]: E1209 09:47:58.738770 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 09:47:58 crc kubenswrapper[4824]: E1209 09:47:58.738830 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 09:48:14.738817181 +0000 UTC m=+51.073321848 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.738048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.738899 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:58Z","lastTransitionTime":"2025-12-09T09:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:58 crc kubenswrapper[4824]: E1209 09:47:58.738923 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 09:47:58 crc kubenswrapper[4824]: E1209 09:47:58.738939 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 09:47:58 crc kubenswrapper[4824]: E1209 09:47:58.738953 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:47:58 crc kubenswrapper[4824]: E1209 09:47:58.738981 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 09:48:14.738974136 +0000 UTC m=+51.073478803 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.740400 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.754933 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.756048 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psr7g\" (UniqueName: \"kubernetes.io/projected/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-kube-api-access-psr7g\") pod \"network-metrics-daemon-5tn52\" (UID: \"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\") " pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.771744 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:58Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.841374 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.841414 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.841425 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.841442 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.841455 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:58Z","lastTransitionTime":"2025-12-09T09:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.909716 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.909846 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:47:58 crc kubenswrapper[4824]: E1209 09:47:58.909921 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.909968 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:47:58 crc kubenswrapper[4824]: E1209 09:47:58.910111 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:47:58 crc kubenswrapper[4824]: E1209 09:47:58.910227 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.944007 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.944051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.944060 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.944078 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:58 crc kubenswrapper[4824]: I1209 09:47:58.944089 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:58Z","lastTransitionTime":"2025-12-09T09:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.047134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.047205 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.047227 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.047249 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.047262 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:59Z","lastTransitionTime":"2025-12-09T09:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.150992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.151046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.151055 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.151302 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.151313 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:59Z","lastTransitionTime":"2025-12-09T09:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.192758 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" event={"ID":"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5","Type":"ContainerStarted","Data":"d37860b2742b9486e06919b4bb4e7e4631ff5905eea837ef17c1d2f8d573da18"} Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.210521 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:59Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.223841 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:59Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.234990 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:59Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.244770 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs\") pod \"network-metrics-daemon-5tn52\" (UID: \"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\") " pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:47:59 crc kubenswrapper[4824]: E1209 09:47:59.244954 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 09:47:59 crc kubenswrapper[4824]: E1209 09:47:59.245041 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs podName:fcbce9ea-3d55-4e09-b5bc-d5bb8c472219 nodeName:}" failed. No retries permitted until 2025-12-09 09:48:00.245017588 +0000 UTC m=+36.579522255 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs") pod "network-metrics-daemon-5tn52" (UID: "fcbce9ea-3d55-4e09-b5bc-d5bb8c472219") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.246566 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:59Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.253633 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.253940 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.254031 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.254126 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.254269 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:59Z","lastTransitionTime":"2025-12-09T09:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.260305 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:59Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.276455 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:59Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.288740 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:59Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.306470 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:59Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.329409 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://102e8ec6dccd87b7bf834dfa6d7c2d18f2e5b9bfe7921d3ecc060056352e3b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://102e8ec6dccd87b7bf834dfa6d7c2d18f2e5b9bfe7921d3ecc060056352e3b27\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:47:56Z\\\",\\\"message\\\":\\\"false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ba175bbe-5cc4-47e6-a32d-57693e1320bd}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 09:47:56.026026 6254 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager/kube-controller-manager]} name:Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ba175bbe-5cc4-47e6-a32d-57693e1320bd}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 09:47:56.026256 6254 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1209 09:47:56.026337 6254 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1209 09:47:56.026374 6254 ovnkube.go:599] Stopped ovnkube\\\\nI1209 09:47:56.026400 6254 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1209 09:47:56.026471 6254 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-vvxck_openshift-ovn-kubernetes(13620d58-6925-48c7-bbc7-f34e92922480)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:59Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.343014 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daad2ae578b4f337ceb3298b012b0262903cebe4e9e9584513cf2f31fbcec3c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37860b2742b9486e06919b4bb4e7e4631ff5905eea837ef17c1d2f8d573da18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gcrqx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:59Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.357829 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.357872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.357891 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.357914 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.357932 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:59Z","lastTransitionTime":"2025-12-09T09:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.359851 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5tn52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5tn52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:59Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.375302 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:59Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.388626 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:59Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.413729 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:59Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.433428 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:59Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.450790 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:59Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.460439 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.460690 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.460770 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.460891 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.460977 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:59Z","lastTransitionTime":"2025-12-09T09:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.466481 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:47:59Z is after 2025-08-24T17:21:41Z" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.563432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.563515 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.563526 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.563546 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.563558 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:59Z","lastTransitionTime":"2025-12-09T09:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.666715 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.667259 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.667275 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.667299 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.667315 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:59Z","lastTransitionTime":"2025-12-09T09:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.769904 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.769960 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.769972 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.769991 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.770004 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:59Z","lastTransitionTime":"2025-12-09T09:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.872406 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.872451 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.872461 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.872478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.872490 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:59Z","lastTransitionTime":"2025-12-09T09:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.910423 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:47:59 crc kubenswrapper[4824]: E1209 09:47:59.910629 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.975607 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.975658 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.975669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.975685 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:47:59 crc kubenswrapper[4824]: I1209 09:47:59.975695 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:47:59Z","lastTransitionTime":"2025-12-09T09:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.079074 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.079119 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.079131 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.079150 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.079163 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:00Z","lastTransitionTime":"2025-12-09T09:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.181972 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.182037 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.182054 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.182077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.182095 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:00Z","lastTransitionTime":"2025-12-09T09:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.257652 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs\") pod \"network-metrics-daemon-5tn52\" (UID: \"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\") " pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:00 crc kubenswrapper[4824]: E1209 09:48:00.257893 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 09:48:00 crc kubenswrapper[4824]: E1209 09:48:00.257976 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs podName:fcbce9ea-3d55-4e09-b5bc-d5bb8c472219 nodeName:}" failed. No retries permitted until 2025-12-09 09:48:02.257951898 +0000 UTC m=+38.592456575 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs") pod "network-metrics-daemon-5tn52" (UID: "fcbce9ea-3d55-4e09-b5bc-d5bb8c472219") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.285634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.285694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.285704 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.285726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.285740 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:00Z","lastTransitionTime":"2025-12-09T09:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.389333 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.389409 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.389428 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.389456 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.389476 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:00Z","lastTransitionTime":"2025-12-09T09:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.492967 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.493066 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.493095 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.493130 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.493154 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:00Z","lastTransitionTime":"2025-12-09T09:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.595834 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.595905 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.595923 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.595950 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.595965 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:00Z","lastTransitionTime":"2025-12-09T09:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.699053 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.699136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.699156 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.699186 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.699207 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:00Z","lastTransitionTime":"2025-12-09T09:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.802439 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.802514 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.802525 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.802546 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.802576 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:00Z","lastTransitionTime":"2025-12-09T09:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.906634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.906691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.906703 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.906724 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.906738 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:00Z","lastTransitionTime":"2025-12-09T09:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.909961 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.910027 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:00 crc kubenswrapper[4824]: I1209 09:48:00.910126 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:00 crc kubenswrapper[4824]: E1209 09:48:00.910327 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:00 crc kubenswrapper[4824]: E1209 09:48:00.910420 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:00 crc kubenswrapper[4824]: E1209 09:48:00.910569 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.008831 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.010008 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.010085 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.010096 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.010114 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.010129 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:01Z","lastTransitionTime":"2025-12-09T09:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.032324 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:01Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.048837 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:01Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.065523 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:01Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.080835 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:01Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.094221 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:01Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.103741 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:01Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.112652 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.112727 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.112740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.112806 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.112821 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:01Z","lastTransitionTime":"2025-12-09T09:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.114250 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:01Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.127764 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:01Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.137457 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:01Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.152428 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:01Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.166420 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:01Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.181956 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:01Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.197034 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:01Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.212331 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:01Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.215518 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.215551 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.215578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.215596 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.215606 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:01Z","lastTransitionTime":"2025-12-09T09:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.233613 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://102e8ec6dccd87b7bf834dfa6d7c2d18f2e5b9bfe7921d3ecc060056352e3b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://102e8ec6dccd87b7bf834dfa6d7c2d18f2e5b9bfe7921d3ecc060056352e3b27\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:47:56Z\\\",\\\"message\\\":\\\"false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ba175bbe-5cc4-47e6-a32d-57693e1320bd}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 09:47:56.026026 6254 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager/kube-controller-manager]} name:Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ba175bbe-5cc4-47e6-a32d-57693e1320bd}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 09:47:56.026256 6254 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1209 09:47:56.026337 6254 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1209 09:47:56.026374 6254 ovnkube.go:599] Stopped ovnkube\\\\nI1209 09:47:56.026400 6254 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1209 09:47:56.026471 6254 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-vvxck_openshift-ovn-kubernetes(13620d58-6925-48c7-bbc7-f34e92922480)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:01Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.247195 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daad2ae578b4f337ceb3298b012b0262903cebe4e9e9584513cf2f31fbcec3c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37860b2742b9486e06919b4bb4e7e4631ff5905eea837ef17c1d2f8d573da18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gcrqx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:01Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.259002 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5tn52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5tn52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:01Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.318590 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.318656 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.318670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.318692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.318706 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:01Z","lastTransitionTime":"2025-12-09T09:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.422019 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.422074 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.422087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.422107 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.422121 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:01Z","lastTransitionTime":"2025-12-09T09:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.525117 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.525154 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.525163 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.525179 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.525190 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:01Z","lastTransitionTime":"2025-12-09T09:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.629687 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.630145 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.630235 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.630333 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.630426 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:01Z","lastTransitionTime":"2025-12-09T09:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.733999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.734069 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.734085 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.734109 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.734127 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:01Z","lastTransitionTime":"2025-12-09T09:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.837774 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.837883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.837901 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.837928 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.838008 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:01Z","lastTransitionTime":"2025-12-09T09:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.910378 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:01 crc kubenswrapper[4824]: E1209 09:48:01.910584 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.941305 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.941391 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.941407 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.941433 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:01 crc kubenswrapper[4824]: I1209 09:48:01.941455 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:01Z","lastTransitionTime":"2025-12-09T09:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.044792 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.044880 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.044895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.044921 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.044939 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:02Z","lastTransitionTime":"2025-12-09T09:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.148467 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.148534 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.148548 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.148565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.148579 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:02Z","lastTransitionTime":"2025-12-09T09:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.251760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.251842 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.251852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.251871 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.251884 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:02Z","lastTransitionTime":"2025-12-09T09:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.281618 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs\") pod \"network-metrics-daemon-5tn52\" (UID: \"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\") " pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:02 crc kubenswrapper[4824]: E1209 09:48:02.281875 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 09:48:02 crc kubenswrapper[4824]: E1209 09:48:02.282018 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs podName:fcbce9ea-3d55-4e09-b5bc-d5bb8c472219 nodeName:}" failed. No retries permitted until 2025-12-09 09:48:06.281992799 +0000 UTC m=+42.616497466 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs") pod "network-metrics-daemon-5tn52" (UID: "fcbce9ea-3d55-4e09-b5bc-d5bb8c472219") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.355012 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.355050 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.355060 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.355076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.355086 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:02Z","lastTransitionTime":"2025-12-09T09:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.458134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.458174 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.458184 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.458202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.458212 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:02Z","lastTransitionTime":"2025-12-09T09:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.561497 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.561547 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.561556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.561575 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.561585 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:02Z","lastTransitionTime":"2025-12-09T09:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.665264 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.665321 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.665340 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.665368 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.665403 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:02Z","lastTransitionTime":"2025-12-09T09:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.770915 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.770992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.771007 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.771028 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.771043 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:02Z","lastTransitionTime":"2025-12-09T09:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.873729 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.873830 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.873845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.873865 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.873877 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:02Z","lastTransitionTime":"2025-12-09T09:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.909872 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:02 crc kubenswrapper[4824]: E1209 09:48:02.910132 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.910169 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.910234 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:02 crc kubenswrapper[4824]: E1209 09:48:02.910411 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:02 crc kubenswrapper[4824]: E1209 09:48:02.910707 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.977238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.977298 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.977314 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.977343 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:02 crc kubenswrapper[4824]: I1209 09:48:02.977365 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:02Z","lastTransitionTime":"2025-12-09T09:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.080538 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.080596 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.080612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.080656 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.080672 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:03Z","lastTransitionTime":"2025-12-09T09:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.184495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.184665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.184696 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.184714 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.184728 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:03Z","lastTransitionTime":"2025-12-09T09:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.288184 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.288227 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.288238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.288254 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.288263 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:03Z","lastTransitionTime":"2025-12-09T09:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.391005 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.391110 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.391124 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.391151 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.391168 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:03Z","lastTransitionTime":"2025-12-09T09:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.494736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.494858 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.494875 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.494896 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.494911 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:03Z","lastTransitionTime":"2025-12-09T09:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.597952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.598020 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.598036 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.598052 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.598064 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:03Z","lastTransitionTime":"2025-12-09T09:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.701584 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.701641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.701658 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.701677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.701690 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:03Z","lastTransitionTime":"2025-12-09T09:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.805296 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.805360 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.805378 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.805401 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.805416 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:03Z","lastTransitionTime":"2025-12-09T09:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.909344 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.909452 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.909486 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.909515 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.909875 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.909530 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:03Z","lastTransitionTime":"2025-12-09T09:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:03 crc kubenswrapper[4824]: E1209 09:48:03.911230 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.928861 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:03Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.942415 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:03Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.956836 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:03Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.979389 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:03Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:03 crc kubenswrapper[4824]: I1209 09:48:03.994899 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:03Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.009952 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:04Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.019223 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.019299 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.019313 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.019352 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.019366 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:04Z","lastTransitionTime":"2025-12-09T09:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.027023 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:04Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.042476 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:04Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.056266 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:04Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.070756 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:04Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.085673 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:04Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.100443 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:04Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.113417 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5tn52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5tn52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:04Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.121960 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.122015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.122027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.122049 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.122062 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:04Z","lastTransitionTime":"2025-12-09T09:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.128882 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:04Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.146142 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:04Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.167309 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://102e8ec6dccd87b7bf834dfa6d7c2d18f2e5b9bfe7921d3ecc060056352e3b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://102e8ec6dccd87b7bf834dfa6d7c2d18f2e5b9bfe7921d3ecc060056352e3b27\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:47:56Z\\\",\\\"message\\\":\\\"false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ba175bbe-5cc4-47e6-a32d-57693e1320bd}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 09:47:56.026026 6254 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager/kube-controller-manager]} name:Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ba175bbe-5cc4-47e6-a32d-57693e1320bd}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 09:47:56.026256 6254 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1209 09:47:56.026337 6254 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1209 09:47:56.026374 6254 ovnkube.go:599] Stopped ovnkube\\\\nI1209 09:47:56.026400 6254 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1209 09:47:56.026471 6254 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-vvxck_openshift-ovn-kubernetes(13620d58-6925-48c7-bbc7-f34e92922480)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:04Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.181407 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daad2ae578b4f337ceb3298b012b0262903cebe4e9e9584513cf2f31fbcec3c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37860b2742b9486e06919b4bb4e7e4631ff5905eea837ef17c1d2f8d573da18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gcrqx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:04Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.223804 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.223841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.223853 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.223868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.223879 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:04Z","lastTransitionTime":"2025-12-09T09:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.326811 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.326888 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.326911 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.326942 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.326964 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:04Z","lastTransitionTime":"2025-12-09T09:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.435059 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.435128 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.435141 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.435162 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.435172 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:04Z","lastTransitionTime":"2025-12-09T09:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.537513 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.537576 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.537593 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.537616 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.537634 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:04Z","lastTransitionTime":"2025-12-09T09:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.639812 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.639848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.639858 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.639873 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.639884 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:04Z","lastTransitionTime":"2025-12-09T09:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.742881 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.742921 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.742934 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.742950 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.742961 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:04Z","lastTransitionTime":"2025-12-09T09:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.846356 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.846416 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.846431 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.846454 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.846472 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:04Z","lastTransitionTime":"2025-12-09T09:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.910376 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:04 crc kubenswrapper[4824]: E1209 09:48:04.910539 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.910996 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.911047 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:04 crc kubenswrapper[4824]: E1209 09:48:04.911377 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:04 crc kubenswrapper[4824]: E1209 09:48:04.911525 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.949300 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.949663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.949729 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.949847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:04 crc kubenswrapper[4824]: I1209 09:48:04.949910 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:04Z","lastTransitionTime":"2025-12-09T09:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.053745 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.053864 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.053886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.053918 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.053936 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:05Z","lastTransitionTime":"2025-12-09T09:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.157592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.157653 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.157667 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.157689 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.157701 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:05Z","lastTransitionTime":"2025-12-09T09:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.262082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.262457 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.262544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.262634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.262721 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:05Z","lastTransitionTime":"2025-12-09T09:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.366604 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.366664 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.366676 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.366697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.366710 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:05Z","lastTransitionTime":"2025-12-09T09:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.470031 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.470121 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.470135 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.470158 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.470176 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:05Z","lastTransitionTime":"2025-12-09T09:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.574731 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.575220 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.575387 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.575531 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.575670 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:05Z","lastTransitionTime":"2025-12-09T09:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.679543 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.680003 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.680170 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.680408 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.680550 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:05Z","lastTransitionTime":"2025-12-09T09:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.784558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.784647 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.784673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.784709 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.784742 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:05Z","lastTransitionTime":"2025-12-09T09:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.888010 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.888051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.888063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.888081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.888099 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:05Z","lastTransitionTime":"2025-12-09T09:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.910438 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:05 crc kubenswrapper[4824]: E1209 09:48:05.910676 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.991832 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.991903 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.991914 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.991932 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:05 crc kubenswrapper[4824]: I1209 09:48:05.991944 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:05Z","lastTransitionTime":"2025-12-09T09:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.096148 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.096217 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.096233 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.096258 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.096274 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:06Z","lastTransitionTime":"2025-12-09T09:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.199202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.199266 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.199281 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.199301 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.199316 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:06Z","lastTransitionTime":"2025-12-09T09:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.302388 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.302438 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.302450 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.302470 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.302483 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:06Z","lastTransitionTime":"2025-12-09T09:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.328412 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs\") pod \"network-metrics-daemon-5tn52\" (UID: \"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\") " pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:06 crc kubenswrapper[4824]: E1209 09:48:06.328619 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 09:48:06 crc kubenswrapper[4824]: E1209 09:48:06.328748 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs podName:fcbce9ea-3d55-4e09-b5bc-d5bb8c472219 nodeName:}" failed. No retries permitted until 2025-12-09 09:48:14.328716479 +0000 UTC m=+50.663221316 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs") pod "network-metrics-daemon-5tn52" (UID: "fcbce9ea-3d55-4e09-b5bc-d5bb8c472219") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.405453 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.405507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.405519 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.405544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.405558 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:06Z","lastTransitionTime":"2025-12-09T09:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.508815 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.508896 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.508917 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.508944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.508966 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:06Z","lastTransitionTime":"2025-12-09T09:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.612161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.612235 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.612259 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.612290 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.612312 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:06Z","lastTransitionTime":"2025-12-09T09:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.717800 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.717876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.717888 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.717911 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.717924 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:06Z","lastTransitionTime":"2025-12-09T09:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.825851 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.825920 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.825935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.825961 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.825977 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:06Z","lastTransitionTime":"2025-12-09T09:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.909824 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.909897 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.909949 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:06 crc kubenswrapper[4824]: E1209 09:48:06.910002 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:06 crc kubenswrapper[4824]: E1209 09:48:06.910115 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:06 crc kubenswrapper[4824]: E1209 09:48:06.910228 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.930031 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.930084 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.930100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.930121 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:06 crc kubenswrapper[4824]: I1209 09:48:06.930133 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:06Z","lastTransitionTime":"2025-12-09T09:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.033382 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.033436 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.033450 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.033468 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.033481 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:07Z","lastTransitionTime":"2025-12-09T09:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.135678 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.135735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.135755 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.135802 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.135820 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:07Z","lastTransitionTime":"2025-12-09T09:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.238618 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.238683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.238699 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.238726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.238747 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:07Z","lastTransitionTime":"2025-12-09T09:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.345334 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.345376 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.345388 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.345408 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.345420 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:07Z","lastTransitionTime":"2025-12-09T09:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.447564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.448363 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.448383 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.448405 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.448420 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:07Z","lastTransitionTime":"2025-12-09T09:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.551669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.551710 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.551720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.551737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.551749 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:07Z","lastTransitionTime":"2025-12-09T09:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.654823 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.654868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.654878 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.654899 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.654910 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:07Z","lastTransitionTime":"2025-12-09T09:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.757401 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.757707 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.758002 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.758215 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.758305 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:07Z","lastTransitionTime":"2025-12-09T09:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.767387 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.767483 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.767501 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.767943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.767971 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:07Z","lastTransitionTime":"2025-12-09T09:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:07 crc kubenswrapper[4824]: E1209 09:48:07.780662 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:07Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.784148 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.784184 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.784196 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.784212 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.784223 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:07Z","lastTransitionTime":"2025-12-09T09:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:07 crc kubenswrapper[4824]: E1209 09:48:07.798391 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:07Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.802758 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.802847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.802859 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.802879 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.802892 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:07Z","lastTransitionTime":"2025-12-09T09:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:07 crc kubenswrapper[4824]: E1209 09:48:07.817267 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:07Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.822557 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.822615 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.822627 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.822654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.822674 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:07Z","lastTransitionTime":"2025-12-09T09:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:07 crc kubenswrapper[4824]: E1209 09:48:07.836670 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:07Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.842067 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.842374 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.842451 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.842534 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.842654 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:07Z","lastTransitionTime":"2025-12-09T09:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:07 crc kubenswrapper[4824]: E1209 09:48:07.857558 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:07Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:07 crc kubenswrapper[4824]: E1209 09:48:07.858403 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.861921 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.861980 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.861991 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.862012 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.862022 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:07Z","lastTransitionTime":"2025-12-09T09:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.909967 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:07 crc kubenswrapper[4824]: E1209 09:48:07.910110 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.963443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.963484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.963492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.963507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:07 crc kubenswrapper[4824]: I1209 09:48:07.963516 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:07Z","lastTransitionTime":"2025-12-09T09:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.065270 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.065308 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.065316 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.065330 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.065339 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:08Z","lastTransitionTime":"2025-12-09T09:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.167233 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.167290 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.167305 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.167326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.167340 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:08Z","lastTransitionTime":"2025-12-09T09:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.270600 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.270657 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.270669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.270688 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.270716 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:08Z","lastTransitionTime":"2025-12-09T09:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.373607 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.373652 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.373662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.373678 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.373687 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:08Z","lastTransitionTime":"2025-12-09T09:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.476880 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.476934 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.476944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.476960 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.476971 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:08Z","lastTransitionTime":"2025-12-09T09:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.579822 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.579860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.579872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.579890 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.579903 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:08Z","lastTransitionTime":"2025-12-09T09:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.682989 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.683515 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.683830 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.684043 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.684212 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:08Z","lastTransitionTime":"2025-12-09T09:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.787215 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.787291 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.787307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.787332 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.787350 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:08Z","lastTransitionTime":"2025-12-09T09:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.890390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.890450 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.890461 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.890500 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.890512 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:08Z","lastTransitionTime":"2025-12-09T09:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.910030 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.910121 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.910178 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:08 crc kubenswrapper[4824]: E1209 09:48:08.910253 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:08 crc kubenswrapper[4824]: E1209 09:48:08.910318 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:08 crc kubenswrapper[4824]: E1209 09:48:08.910511 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.994162 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.994227 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.994249 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.994278 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:08 crc kubenswrapper[4824]: I1209 09:48:08.994299 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:08Z","lastTransitionTime":"2025-12-09T09:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.098395 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.098440 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.098457 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.098473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.098484 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:09Z","lastTransitionTime":"2025-12-09T09:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.201504 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.201562 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.201574 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.201597 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.201672 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:09Z","lastTransitionTime":"2025-12-09T09:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.305480 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.305570 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.305585 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.305609 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.305628 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:09Z","lastTransitionTime":"2025-12-09T09:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.409093 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.409141 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.409152 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.409170 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.409183 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:09Z","lastTransitionTime":"2025-12-09T09:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.512514 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.512563 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.512575 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.512594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.512606 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:09Z","lastTransitionTime":"2025-12-09T09:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.615921 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.615970 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.615979 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.615995 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.616005 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:09Z","lastTransitionTime":"2025-12-09T09:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.719517 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.719560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.719569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.719589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.719600 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:09Z","lastTransitionTime":"2025-12-09T09:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.823411 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.823475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.823491 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.823511 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.823522 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:09Z","lastTransitionTime":"2025-12-09T09:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.909950 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:09 crc kubenswrapper[4824]: E1209 09:48:09.910191 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.925795 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.926097 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.926168 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.926236 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:09 crc kubenswrapper[4824]: I1209 09:48:09.926307 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:09Z","lastTransitionTime":"2025-12-09T09:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.029590 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.029971 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.030056 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.030138 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.030213 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:10Z","lastTransitionTime":"2025-12-09T09:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.133285 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.133342 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.133353 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.133371 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.133386 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:10Z","lastTransitionTime":"2025-12-09T09:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.236027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.236068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.236079 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.236096 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.236109 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:10Z","lastTransitionTime":"2025-12-09T09:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.339284 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.339357 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.339369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.339389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.339402 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:10Z","lastTransitionTime":"2025-12-09T09:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.443388 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.443808 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.443896 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.443969 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.444033 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:10Z","lastTransitionTime":"2025-12-09T09:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.547073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.547112 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.547124 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.547142 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.547153 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:10Z","lastTransitionTime":"2025-12-09T09:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.650440 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.650504 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.650518 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.650543 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.650556 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:10Z","lastTransitionTime":"2025-12-09T09:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.754383 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.754892 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.755052 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.755208 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.755307 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:10Z","lastTransitionTime":"2025-12-09T09:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.859006 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.859442 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.859547 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.859633 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.859707 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:10Z","lastTransitionTime":"2025-12-09T09:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.910223 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.910223 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.910397 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:10 crc kubenswrapper[4824]: E1209 09:48:10.910967 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:10 crc kubenswrapper[4824]: E1209 09:48:10.911002 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:10 crc kubenswrapper[4824]: E1209 09:48:10.911076 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.911444 4824 scope.go:117] "RemoveContainer" containerID="102e8ec6dccd87b7bf834dfa6d7c2d18f2e5b9bfe7921d3ecc060056352e3b27" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.964316 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.964562 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.964574 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.964596 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:10 crc kubenswrapper[4824]: I1209 09:48:10.964609 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:10Z","lastTransitionTime":"2025-12-09T09:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.068175 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.068215 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.068229 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.068254 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.068270 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:11Z","lastTransitionTime":"2025-12-09T09:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.172601 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.172669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.172680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.172701 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.172713 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:11Z","lastTransitionTime":"2025-12-09T09:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.237043 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vvxck_13620d58-6925-48c7-bbc7-f34e92922480/ovnkube-controller/1.log" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.240493 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerStarted","Data":"352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34"} Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.241134 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.267368 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:11Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.275314 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.275376 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.275388 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.275408 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.275421 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:11Z","lastTransitionTime":"2025-12-09T09:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.299509 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:11Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.325273 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:11Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.342982 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:11Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.366148 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:11Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.377709 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.377764 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.377775 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.377826 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.377845 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:11Z","lastTransitionTime":"2025-12-09T09:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.387408 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:11Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.406319 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:11Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.421370 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:11Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.434521 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:11Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.449163 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:11Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.466116 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:11Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.479964 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:11Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.480748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.480829 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.480848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.480867 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.480881 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:11Z","lastTransitionTime":"2025-12-09T09:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.493578 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:11Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.512590 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:11Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.532207 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://102e8ec6dccd87b7bf834dfa6d7c2d18f2e5b9bfe7921d3ecc060056352e3b27\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:47:56Z\\\",\\\"message\\\":\\\"false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ba175bbe-5cc4-47e6-a32d-57693e1320bd}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 09:47:56.026026 6254 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager/kube-controller-manager]} name:Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ba175bbe-5cc4-47e6-a32d-57693e1320bd}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 09:47:56.026256 6254 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1209 09:47:56.026337 6254 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1209 09:47:56.026374 6254 ovnkube.go:599] Stopped ovnkube\\\\nI1209 09:47:56.026400 6254 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1209 09:47:56.026471 6254 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:48:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:11Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.544058 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daad2ae578b4f337ceb3298b012b0262903cebe4e9e9584513cf2f31fbcec3c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37860b2742b9486e06919b4bb4e7e4631ff5905eea837ef17c1d2f8d573da18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gcrqx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:11Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.561178 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5tn52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5tn52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:11Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.583413 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.583458 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.583469 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.583488 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.583500 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:11Z","lastTransitionTime":"2025-12-09T09:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.686216 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.686304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.686329 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.686360 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.686385 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:11Z","lastTransitionTime":"2025-12-09T09:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.788975 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.789035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.789047 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.789066 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.789080 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:11Z","lastTransitionTime":"2025-12-09T09:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.891797 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.891833 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.891849 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.891865 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.891876 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:11Z","lastTransitionTime":"2025-12-09T09:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.909741 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:11 crc kubenswrapper[4824]: E1209 09:48:11.910013 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.994234 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.994288 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.994299 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.994314 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:11 crc kubenswrapper[4824]: I1209 09:48:11.994326 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:11Z","lastTransitionTime":"2025-12-09T09:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.097045 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.097555 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.097681 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.097884 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.098019 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:12Z","lastTransitionTime":"2025-12-09T09:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.200720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.200828 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.200852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.200885 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.200904 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:12Z","lastTransitionTime":"2025-12-09T09:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.246678 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vvxck_13620d58-6925-48c7-bbc7-f34e92922480/ovnkube-controller/2.log" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.248109 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vvxck_13620d58-6925-48c7-bbc7-f34e92922480/ovnkube-controller/1.log" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.252174 4824 generic.go:334] "Generic (PLEG): container finished" podID="13620d58-6925-48c7-bbc7-f34e92922480" containerID="352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34" exitCode=1 Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.252238 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerDied","Data":"352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34"} Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.252432 4824 scope.go:117] "RemoveContainer" containerID="102e8ec6dccd87b7bf834dfa6d7c2d18f2e5b9bfe7921d3ecc060056352e3b27" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.253013 4824 scope.go:117] "RemoveContainer" containerID="352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34" Dec 09 09:48:12 crc kubenswrapper[4824]: E1209 09:48:12.253354 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vvxck_openshift-ovn-kubernetes(13620d58-6925-48c7-bbc7-f34e92922480)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" podUID="13620d58-6925-48c7-bbc7-f34e92922480" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.272551 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:12Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.288228 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:12Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.302230 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:12Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.304948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.304980 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.304992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.305014 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.305027 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:12Z","lastTransitionTime":"2025-12-09T09:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.315025 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daad2ae578b4f337ceb3298b012b0262903cebe4e9e9584513cf2f31fbcec3c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37860b2742b9486e06919b4bb4e7e4631ff5905eea837ef17c1d2f8d573da18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gcrqx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:12Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.326452 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5tn52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5tn52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:12Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.339506 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:12Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.356123 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:12Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.378640 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://102e8ec6dccd87b7bf834dfa6d7c2d18f2e5b9bfe7921d3ecc060056352e3b27\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:47:56Z\\\",\\\"message\\\":\\\"false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ba175bbe-5cc4-47e6-a32d-57693e1320bd}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 09:47:56.026026 6254 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager/kube-controller-manager]} name:Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ba175bbe-5cc4-47e6-a32d-57693e1320bd}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 09:47:56.026256 6254 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1209 09:47:56.026337 6254 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1209 09:47:56.026374 6254 ovnkube.go:599] Stopped ovnkube\\\\nI1209 09:47:56.026400 6254 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1209 09:47:56.026471 6254 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:48:11Z\\\",\\\"message\\\":\\\"ncer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI1209 09:48:11.888995 6487 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 2.013563ms\\\\nI1209 09:48:11.889004 6487 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1209 09:48:11.889013 6487 services_controller.go:356] Processing sync for service openshift-operator-lifecycle-manager/packageserver-service for network=default\\\\nI1209 09:48:11.889016 6487 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1209 09:48:11.889010 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": fail\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:48:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:12Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.391911 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:12Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.406299 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:12Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.408162 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.408191 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.408202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.408218 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.408231 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:12Z","lastTransitionTime":"2025-12-09T09:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.418368 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:12Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.429331 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:12Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.446941 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:12Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.459220 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:12Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.472097 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:12Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.487325 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:12Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.500751 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:12Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.510443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.510472 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.510480 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.510494 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.510503 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:12Z","lastTransitionTime":"2025-12-09T09:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.613257 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.613299 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.613310 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.613327 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.613337 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:12Z","lastTransitionTime":"2025-12-09T09:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.716319 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.716371 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.716381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.716397 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.716407 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:12Z","lastTransitionTime":"2025-12-09T09:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.820040 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.820348 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.820422 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.820490 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.820554 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:12Z","lastTransitionTime":"2025-12-09T09:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.909656 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.909710 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.909680 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:12 crc kubenswrapper[4824]: E1209 09:48:12.909875 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:12 crc kubenswrapper[4824]: E1209 09:48:12.910025 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:12 crc kubenswrapper[4824]: E1209 09:48:12.910191 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.923025 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.923139 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.923156 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.923175 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:12 crc kubenswrapper[4824]: I1209 09:48:12.923188 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:12Z","lastTransitionTime":"2025-12-09T09:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.026415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.026449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.026457 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.026471 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.026481 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:13Z","lastTransitionTime":"2025-12-09T09:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.129889 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.129967 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.129980 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.129998 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.130011 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:13Z","lastTransitionTime":"2025-12-09T09:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.232697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.232752 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.232764 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.232807 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.232820 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:13Z","lastTransitionTime":"2025-12-09T09:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.258839 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vvxck_13620d58-6925-48c7-bbc7-f34e92922480/ovnkube-controller/2.log" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.264252 4824 scope.go:117] "RemoveContainer" containerID="352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34" Dec 09 09:48:13 crc kubenswrapper[4824]: E1209 09:48:13.264443 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vvxck_openshift-ovn-kubernetes(13620d58-6925-48c7-bbc7-f34e92922480)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" podUID="13620d58-6925-48c7-bbc7-f34e92922480" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.289593 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:48:11Z\\\",\\\"message\\\":\\\"ncer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI1209 09:48:11.888995 6487 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 2.013563ms\\\\nI1209 09:48:11.889004 6487 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1209 09:48:11.889013 6487 services_controller.go:356] Processing sync for service openshift-operator-lifecycle-manager/packageserver-service for network=default\\\\nI1209 09:48:11.889016 6487 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1209 09:48:11.889010 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": fail\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:48:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vvxck_openshift-ovn-kubernetes(13620d58-6925-48c7-bbc7-f34e92922480)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:13Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.303364 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daad2ae578b4f337ceb3298b012b0262903cebe4e9e9584513cf2f31fbcec3c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37860b2742b9486e06919b4bb4e7e4631ff5905eea837ef17c1d2f8d573da18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gcrqx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:13Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.316110 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5tn52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5tn52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:13Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.328718 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:13Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.335448 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.335481 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.335492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.335511 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.335523 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:13Z","lastTransitionTime":"2025-12-09T09:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.344056 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:13Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.358639 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:13Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.372707 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:13Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.385149 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:13Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.397990 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:13Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.411506 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:13Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.430898 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:13Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.438760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.438821 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.438835 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.438857 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.438872 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:13Z","lastTransitionTime":"2025-12-09T09:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.443367 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:13Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.493298 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:13Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.515639 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:13Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.532517 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:13Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.541364 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.541417 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.541431 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.541451 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.541464 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:13Z","lastTransitionTime":"2025-12-09T09:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.549088 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:13Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.563309 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:13Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.644092 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.644162 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.644179 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.644204 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.644220 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:13Z","lastTransitionTime":"2025-12-09T09:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.747157 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.747279 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.747303 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.747336 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.747357 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:13Z","lastTransitionTime":"2025-12-09T09:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.853881 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.853956 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.853981 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.854010 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.854037 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:13Z","lastTransitionTime":"2025-12-09T09:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.910013 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:13 crc kubenswrapper[4824]: E1209 09:48:13.910204 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.925523 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daad2ae578b4f337ceb3298b012b0262903cebe4e9e9584513cf2f31fbcec3c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37860b2742b9486e06919b4bb4e7e4631ff5905eea837ef17c1d2f8d573da18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gcrqx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:13Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.940913 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5tn52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5tn52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:13Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.957912 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.958276 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.958303 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.958325 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.958341 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:13Z","lastTransitionTime":"2025-12-09T09:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.959322 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:13Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:13 crc kubenswrapper[4824]: I1209 09:48:13.977745 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:13Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.006155 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:48:11Z\\\",\\\"message\\\":\\\"ncer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI1209 09:48:11.888995 6487 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 2.013563ms\\\\nI1209 09:48:11.889004 6487 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1209 09:48:11.889013 6487 services_controller.go:356] Processing sync for service openshift-operator-lifecycle-manager/packageserver-service for network=default\\\\nI1209 09:48:11.889016 6487 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1209 09:48:11.889010 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": fail\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:48:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vvxck_openshift-ovn-kubernetes(13620d58-6925-48c7-bbc7-f34e92922480)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:14Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.028182 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:14Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.046637 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:14Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.060773 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:14Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.061152 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.061179 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.061193 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.061213 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.061227 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:14Z","lastTransitionTime":"2025-12-09T09:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.075667 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:14Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.101613 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:14Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.118468 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:14Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.132924 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:14Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.148175 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:14Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.164206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.164259 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.164389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.164424 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.164440 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:14Z","lastTransitionTime":"2025-12-09T09:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.165091 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:14Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.181678 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:14Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.199054 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:14Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.216035 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:14Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.267517 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.267562 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.267573 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.267592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.267603 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:14Z","lastTransitionTime":"2025-12-09T09:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.370218 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.370314 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.370327 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.370354 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.370374 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:14Z","lastTransitionTime":"2025-12-09T09:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.413422 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs\") pod \"network-metrics-daemon-5tn52\" (UID: \"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\") " pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:14 crc kubenswrapper[4824]: E1209 09:48:14.413586 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 09:48:14 crc kubenswrapper[4824]: E1209 09:48:14.413658 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs podName:fcbce9ea-3d55-4e09-b5bc-d5bb8c472219 nodeName:}" failed. No retries permitted until 2025-12-09 09:48:30.413638241 +0000 UTC m=+66.748142908 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs") pod "network-metrics-daemon-5tn52" (UID: "fcbce9ea-3d55-4e09-b5bc-d5bb8c472219") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.472882 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.472933 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.472944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.472962 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.472974 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:14Z","lastTransitionTime":"2025-12-09T09:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.576202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.576275 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.576290 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.576317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.576329 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:14Z","lastTransitionTime":"2025-12-09T09:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.679641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.679700 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.679711 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.679734 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.679747 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:14Z","lastTransitionTime":"2025-12-09T09:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.716264 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:48:14 crc kubenswrapper[4824]: E1209 09:48:14.716518 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:48:46.716474709 +0000 UTC m=+83.050979416 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.782472 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.782545 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.782557 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.782578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.782589 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:14Z","lastTransitionTime":"2025-12-09T09:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.817646 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.817702 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.817736 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.817768 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:14 crc kubenswrapper[4824]: E1209 09:48:14.817823 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 09:48:14 crc kubenswrapper[4824]: E1209 09:48:14.817891 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 09:48:14 crc kubenswrapper[4824]: E1209 09:48:14.817908 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 09:48:14 crc kubenswrapper[4824]: E1209 09:48:14.817919 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:48:14 crc kubenswrapper[4824]: E1209 09:48:14.817992 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 09:48:14 crc kubenswrapper[4824]: E1209 09:48:14.818004 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 09:48:14 crc kubenswrapper[4824]: E1209 09:48:14.818013 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:48:14 crc kubenswrapper[4824]: E1209 09:48:14.818077 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 09:48:14 crc kubenswrapper[4824]: E1209 09:48:14.817896 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 09:48:46.81787328 +0000 UTC m=+83.152377947 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 09:48:14 crc kubenswrapper[4824]: E1209 09:48:14.818197 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 09:48:46.818140548 +0000 UTC m=+83.152645225 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:48:14 crc kubenswrapper[4824]: E1209 09:48:14.818225 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 09:48:46.818214371 +0000 UTC m=+83.152719158 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:48:14 crc kubenswrapper[4824]: E1209 09:48:14.818250 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 09:48:46.818240181 +0000 UTC m=+83.152744948 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.885389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.885431 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.885440 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.885455 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.885465 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:14Z","lastTransitionTime":"2025-12-09T09:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.910064 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.910064 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.910084 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:14 crc kubenswrapper[4824]: E1209 09:48:14.910366 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:14 crc kubenswrapper[4824]: E1209 09:48:14.910220 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:14 crc kubenswrapper[4824]: E1209 09:48:14.910510 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.988123 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.988191 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.988208 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.988231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:14 crc kubenswrapper[4824]: I1209 09:48:14.988249 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:14Z","lastTransitionTime":"2025-12-09T09:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.091689 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.091971 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.091987 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.092013 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.092030 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:15Z","lastTransitionTime":"2025-12-09T09:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.194989 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.195090 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.195116 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.195152 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.195176 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:15Z","lastTransitionTime":"2025-12-09T09:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.297889 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.297943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.297955 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.297973 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.297987 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:15Z","lastTransitionTime":"2025-12-09T09:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.400697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.400758 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.400771 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.400808 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.400822 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:15Z","lastTransitionTime":"2025-12-09T09:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.505583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.505978 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.505989 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.506011 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.506024 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:15Z","lastTransitionTime":"2025-12-09T09:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.609154 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.609196 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.609206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.609222 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.609234 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:15Z","lastTransitionTime":"2025-12-09T09:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.712295 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.712337 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.712346 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.712363 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.712372 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:15Z","lastTransitionTime":"2025-12-09T09:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.814259 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.814309 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.814322 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.814340 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.814353 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:15Z","lastTransitionTime":"2025-12-09T09:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.909766 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:15 crc kubenswrapper[4824]: E1209 09:48:15.910000 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.916158 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.916204 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.916219 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.916237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:15 crc kubenswrapper[4824]: I1209 09:48:15.916252 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:15Z","lastTransitionTime":"2025-12-09T09:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.018403 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.018450 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.018460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.018484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.018496 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:16Z","lastTransitionTime":"2025-12-09T09:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.060569 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.072952 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:16Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.074952 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.087866 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:16Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.103012 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:16Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.121107 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.121157 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.121168 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.121189 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.121220 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:16Z","lastTransitionTime":"2025-12-09T09:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.123932 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:48:11Z\\\",\\\"message\\\":\\\"ncer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI1209 09:48:11.888995 6487 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 2.013563ms\\\\nI1209 09:48:11.889004 6487 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1209 09:48:11.889013 6487 services_controller.go:356] Processing sync for service openshift-operator-lifecycle-manager/packageserver-service for network=default\\\\nI1209 09:48:11.889016 6487 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1209 09:48:11.889010 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": fail\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:48:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vvxck_openshift-ovn-kubernetes(13620d58-6925-48c7-bbc7-f34e92922480)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:16Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.137332 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daad2ae578b4f337ceb3298b012b0262903cebe4e9e9584513cf2f31fbcec3c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37860b2742b9486e06919b4bb4e7e4631ff5905eea837ef17c1d2f8d573da18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gcrqx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:16Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.147708 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5tn52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5tn52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:16Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.159549 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:16Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.172390 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:16Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.188711 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:16Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.203291 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:16Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.219862 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:16Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.224166 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.224208 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.224238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.224256 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.224267 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:16Z","lastTransitionTime":"2025-12-09T09:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.231059 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:16Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.243213 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:16Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.264637 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:16Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.275575 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:16Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.292705 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:16Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.307464 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:16Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.326389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.326447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.326457 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.326474 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.326486 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:16Z","lastTransitionTime":"2025-12-09T09:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.428962 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.429009 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.429019 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.429038 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.429049 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:16Z","lastTransitionTime":"2025-12-09T09:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.532104 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.532148 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.532158 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.532175 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.532186 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:16Z","lastTransitionTime":"2025-12-09T09:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.634598 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.634654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.634666 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.634693 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.634705 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:16Z","lastTransitionTime":"2025-12-09T09:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.738572 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.738643 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.738656 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.738678 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.738693 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:16Z","lastTransitionTime":"2025-12-09T09:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.841621 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.841710 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.841725 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.841748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.841762 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:16Z","lastTransitionTime":"2025-12-09T09:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.910391 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.910537 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.910395 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:16 crc kubenswrapper[4824]: E1209 09:48:16.910610 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:16 crc kubenswrapper[4824]: E1209 09:48:16.910712 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:16 crc kubenswrapper[4824]: E1209 09:48:16.910806 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.944848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.944882 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.944893 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.944910 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:16 crc kubenswrapper[4824]: I1209 09:48:16.944920 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:16Z","lastTransitionTime":"2025-12-09T09:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.047707 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.047755 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.047769 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.047823 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.047837 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:17Z","lastTransitionTime":"2025-12-09T09:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.150597 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.150641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.150652 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.150673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.150684 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:17Z","lastTransitionTime":"2025-12-09T09:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.253295 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.253379 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.253391 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.253416 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.253433 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:17Z","lastTransitionTime":"2025-12-09T09:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.356466 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.356500 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.356513 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.356529 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.356540 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:17Z","lastTransitionTime":"2025-12-09T09:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.459179 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.459229 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.459242 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.459293 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.459307 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:17Z","lastTransitionTime":"2025-12-09T09:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.561680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.561732 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.561742 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.561757 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.561767 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:17Z","lastTransitionTime":"2025-12-09T09:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.664832 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.664883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.664895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.664916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.664929 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:17Z","lastTransitionTime":"2025-12-09T09:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.767865 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.767921 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.767936 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.767959 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.767970 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:17Z","lastTransitionTime":"2025-12-09T09:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.871172 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.871264 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.871307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.871343 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.871368 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:17Z","lastTransitionTime":"2025-12-09T09:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.909651 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:17 crc kubenswrapper[4824]: E1209 09:48:17.909969 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.918076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.918111 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.918121 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.918132 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.918143 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:17Z","lastTransitionTime":"2025-12-09T09:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:17 crc kubenswrapper[4824]: E1209 09:48:17.930006 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:17Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.933378 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.933421 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.933430 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.933447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.933458 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:17Z","lastTransitionTime":"2025-12-09T09:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:17 crc kubenswrapper[4824]: E1209 09:48:17.950992 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:17Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.954614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.954663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.954675 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.954694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.954706 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:17Z","lastTransitionTime":"2025-12-09T09:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:17 crc kubenswrapper[4824]: E1209 09:48:17.968656 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:17Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.972218 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.972262 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.972272 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.972287 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.972298 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:17Z","lastTransitionTime":"2025-12-09T09:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:17 crc kubenswrapper[4824]: E1209 09:48:17.984037 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:17Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.987624 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.987700 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.987723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.987740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:17 crc kubenswrapper[4824]: I1209 09:48:17.987751 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:17Z","lastTransitionTime":"2025-12-09T09:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:17 crc kubenswrapper[4824]: E1209 09:48:17.999084 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:17Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:17 crc kubenswrapper[4824]: E1209 09:48:17.999196 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.001036 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.001088 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.001118 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.001133 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.001168 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:18Z","lastTransitionTime":"2025-12-09T09:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.103985 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.104038 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.104050 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.104067 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.104080 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:18Z","lastTransitionTime":"2025-12-09T09:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.206822 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.206867 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.206877 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.206890 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.206903 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:18Z","lastTransitionTime":"2025-12-09T09:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.309369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.309409 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.309419 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.309435 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.309448 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:18Z","lastTransitionTime":"2025-12-09T09:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.413848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.413926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.413945 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.413974 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.413994 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:18Z","lastTransitionTime":"2025-12-09T09:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.517392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.517458 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.517476 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.517501 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.517521 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:18Z","lastTransitionTime":"2025-12-09T09:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.621246 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.621299 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.621313 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.621332 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.621349 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:18Z","lastTransitionTime":"2025-12-09T09:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.724063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.724471 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.724630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.724735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.724853 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:18Z","lastTransitionTime":"2025-12-09T09:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.827663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.827713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.827726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.827745 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.827762 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:18Z","lastTransitionTime":"2025-12-09T09:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.910215 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:18 crc kubenswrapper[4824]: E1209 09:48:18.910757 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.910311 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:18 crc kubenswrapper[4824]: E1209 09:48:18.910998 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.910301 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:18 crc kubenswrapper[4824]: E1209 09:48:18.911201 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.931090 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.931415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.931484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.931601 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:18 crc kubenswrapper[4824]: I1209 09:48:18.931663 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:18Z","lastTransitionTime":"2025-12-09T09:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.035289 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.035335 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.035344 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.035364 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.035374 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:19Z","lastTransitionTime":"2025-12-09T09:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.139389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.139453 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.139472 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.139502 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.139521 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:19Z","lastTransitionTime":"2025-12-09T09:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.243137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.243187 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.243200 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.243220 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.243237 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:19Z","lastTransitionTime":"2025-12-09T09:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.345926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.345963 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.345972 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.345986 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.345997 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:19Z","lastTransitionTime":"2025-12-09T09:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.448231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.448282 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.448318 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.448336 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.448349 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:19Z","lastTransitionTime":"2025-12-09T09:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.552101 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.552243 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.552592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.552630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.552645 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:19Z","lastTransitionTime":"2025-12-09T09:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.655364 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.655423 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.655436 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.655460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.655476 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:19Z","lastTransitionTime":"2025-12-09T09:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.758345 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.758412 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.758426 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.758446 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.758457 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:19Z","lastTransitionTime":"2025-12-09T09:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.861427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.861476 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.861490 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.861508 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.861523 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:19Z","lastTransitionTime":"2025-12-09T09:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.910068 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:19 crc kubenswrapper[4824]: E1209 09:48:19.910242 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.965955 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.966031 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.966056 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.966088 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:19 crc kubenswrapper[4824]: I1209 09:48:19.966106 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:19Z","lastTransitionTime":"2025-12-09T09:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.069625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.069708 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.069732 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.069768 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.069830 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:20Z","lastTransitionTime":"2025-12-09T09:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.173738 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.174102 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.174292 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.174468 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.174606 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:20Z","lastTransitionTime":"2025-12-09T09:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.276872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.276932 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.276943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.276960 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.276995 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:20Z","lastTransitionTime":"2025-12-09T09:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.379977 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.380448 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.380518 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.380587 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.380650 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:20Z","lastTransitionTime":"2025-12-09T09:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.484896 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.484975 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.484993 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.485017 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.485047 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:20Z","lastTransitionTime":"2025-12-09T09:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.588274 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.588330 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.588342 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.588360 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.588376 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:20Z","lastTransitionTime":"2025-12-09T09:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.692731 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.692859 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.692888 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.692924 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.692948 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:20Z","lastTransitionTime":"2025-12-09T09:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.795835 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.795920 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.795946 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.795984 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.796009 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:20Z","lastTransitionTime":"2025-12-09T09:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.899694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.899770 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.899818 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.899847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.899866 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:20Z","lastTransitionTime":"2025-12-09T09:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.910012 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.910041 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:20 crc kubenswrapper[4824]: E1209 09:48:20.910254 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:20 crc kubenswrapper[4824]: I1209 09:48:20.910055 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:20 crc kubenswrapper[4824]: E1209 09:48:20.910435 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:20 crc kubenswrapper[4824]: E1209 09:48:20.910578 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.004601 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.005064 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.005162 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.005263 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.005353 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:21Z","lastTransitionTime":"2025-12-09T09:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.108389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.108668 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.108731 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.108835 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.108936 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:21Z","lastTransitionTime":"2025-12-09T09:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.212894 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.213347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.213513 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.213702 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.213974 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:21Z","lastTransitionTime":"2025-12-09T09:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.317413 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.317453 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.317462 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.317476 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.317486 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:21Z","lastTransitionTime":"2025-12-09T09:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.420286 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.420343 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.420354 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.420372 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.420384 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:21Z","lastTransitionTime":"2025-12-09T09:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.523459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.523515 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.523525 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.523549 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.523561 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:21Z","lastTransitionTime":"2025-12-09T09:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.626470 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.626531 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.626550 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.626574 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.626593 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:21Z","lastTransitionTime":"2025-12-09T09:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.729896 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.729971 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.729981 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.729998 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.730009 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:21Z","lastTransitionTime":"2025-12-09T09:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.832556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.832886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.832915 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.832944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.832959 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:21Z","lastTransitionTime":"2025-12-09T09:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.910070 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:21 crc kubenswrapper[4824]: E1209 09:48:21.910311 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.935357 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.935422 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.935436 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.935459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:21 crc kubenswrapper[4824]: I1209 09:48:21.935473 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:21Z","lastTransitionTime":"2025-12-09T09:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.038081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.038136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.038148 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.038166 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.038179 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:22Z","lastTransitionTime":"2025-12-09T09:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.140683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.140747 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.140763 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.140829 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.140854 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:22Z","lastTransitionTime":"2025-12-09T09:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.243770 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.243835 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.243847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.243864 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.243879 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:22Z","lastTransitionTime":"2025-12-09T09:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.346829 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.346911 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.346933 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.346962 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.347049 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:22Z","lastTransitionTime":"2025-12-09T09:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.450522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.450612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.450625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.450644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.450656 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:22Z","lastTransitionTime":"2025-12-09T09:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.554609 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.554665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.554677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.554696 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.554709 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:22Z","lastTransitionTime":"2025-12-09T09:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.657976 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.658017 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.658026 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.658045 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.658055 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:22Z","lastTransitionTime":"2025-12-09T09:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.760906 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.760961 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.760973 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.760993 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.761009 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:22Z","lastTransitionTime":"2025-12-09T09:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.863104 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.863136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.863146 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.863160 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.863170 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:22Z","lastTransitionTime":"2025-12-09T09:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.910226 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.910305 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.910323 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:22 crc kubenswrapper[4824]: E1209 09:48:22.910434 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:22 crc kubenswrapper[4824]: E1209 09:48:22.910532 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:22 crc kubenswrapper[4824]: E1209 09:48:22.910666 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.965242 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.965311 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.965325 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.965344 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:22 crc kubenswrapper[4824]: I1209 09:48:22.965355 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:22Z","lastTransitionTime":"2025-12-09T09:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.068376 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.068423 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.068432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.068456 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.068465 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:23Z","lastTransitionTime":"2025-12-09T09:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.171406 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.171464 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.171478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.171501 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.171514 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:23Z","lastTransitionTime":"2025-12-09T09:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.274025 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.274073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.274083 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.274130 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.274143 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:23Z","lastTransitionTime":"2025-12-09T09:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.376928 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.376990 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.377000 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.377018 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.377029 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:23Z","lastTransitionTime":"2025-12-09T09:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.479824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.479876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.479886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.479905 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.479916 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:23Z","lastTransitionTime":"2025-12-09T09:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.582861 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.582988 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.583002 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.583019 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.583054 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:23Z","lastTransitionTime":"2025-12-09T09:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.686586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.686665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.686677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.686697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.686709 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:23Z","lastTransitionTime":"2025-12-09T09:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.789894 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.789954 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.789971 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.789995 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.790014 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:23Z","lastTransitionTime":"2025-12-09T09:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.893012 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.893088 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.893114 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.893145 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.893170 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:23Z","lastTransitionTime":"2025-12-09T09:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.909484 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:23 crc kubenswrapper[4824]: E1209 09:48:23.909674 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.931343 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:23Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.948124 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:23Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.963716 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:23Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.978485 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2dd916-62ee-4dca-964d-fb859475a9d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1e6cfeb997cafb7ec7f6d829e1ff2e2d95e4b343c422194a5290d222ea53a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2add2743e390cf2b91c9b020de38fc6df392fee9e5aa1d59bd1ca9471d9d8d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17601a452dc4369a533806384d58f733e396e255f5741b38004cae98a44cfe55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eb98ce1943d5692f7e6d51befb0c05115ee403be72134f40f651a80bf0a378b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8eb98ce1943d5692f7e6d51befb0c05115ee403be72134f40f651a80bf0a378b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:23Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.996355 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:23Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.996429 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.996820 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.996852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.996886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:23 crc kubenswrapper[4824]: I1209 09:48:23.996911 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:23Z","lastTransitionTime":"2025-12-09T09:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.015768 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:24Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.033389 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:24Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.048426 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:24Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.069982 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:24Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.096389 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:48:11Z\\\",\\\"message\\\":\\\"ncer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI1209 09:48:11.888995 6487 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 2.013563ms\\\\nI1209 09:48:11.889004 6487 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1209 09:48:11.889013 6487 services_controller.go:356] Processing sync for service openshift-operator-lifecycle-manager/packageserver-service for network=default\\\\nI1209 09:48:11.889016 6487 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1209 09:48:11.889010 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": fail\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:48:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vvxck_openshift-ovn-kubernetes(13620d58-6925-48c7-bbc7-f34e92922480)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:24Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.099144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.099195 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.099224 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.099253 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.099271 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:24Z","lastTransitionTime":"2025-12-09T09:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.111211 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daad2ae578b4f337ceb3298b012b0262903cebe4e9e9584513cf2f31fbcec3c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37860b2742b9486e06919b4bb4e7e4631ff5905eea837ef17c1d2f8d573da18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gcrqx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:24Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.124534 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5tn52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5tn52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:24Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.149045 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:24Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.164855 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:24Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.177216 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:24Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.188617 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:24Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.201562 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.201597 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.201607 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.201626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.201638 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:24Z","lastTransitionTime":"2025-12-09T09:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.205243 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:24Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.219320 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:24Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.303571 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.303628 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.303641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.303659 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.303673 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:24Z","lastTransitionTime":"2025-12-09T09:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.406650 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.406712 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.406729 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.406754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.406771 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:24Z","lastTransitionTime":"2025-12-09T09:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.510752 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.510828 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.510841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.510874 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.510890 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:24Z","lastTransitionTime":"2025-12-09T09:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.613840 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.613878 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.613891 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.613908 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.613919 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:24Z","lastTransitionTime":"2025-12-09T09:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.718872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.718917 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.718927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.718941 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.718953 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:24Z","lastTransitionTime":"2025-12-09T09:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.821049 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.821125 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.821146 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.821173 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.821195 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:24Z","lastTransitionTime":"2025-12-09T09:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.909955 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.909998 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:24 crc kubenswrapper[4824]: E1209 09:48:24.910095 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.909965 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:24 crc kubenswrapper[4824]: E1209 09:48:24.910217 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:24 crc kubenswrapper[4824]: E1209 09:48:24.910328 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.923154 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.923215 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.923232 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.923260 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:24 crc kubenswrapper[4824]: I1209 09:48:24.923276 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:24Z","lastTransitionTime":"2025-12-09T09:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.027030 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.027112 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.027138 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.027168 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.027188 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:25Z","lastTransitionTime":"2025-12-09T09:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.129972 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.130044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.130063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.130092 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.130111 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:25Z","lastTransitionTime":"2025-12-09T09:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.233050 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.233133 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.233168 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.233191 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.233209 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:25Z","lastTransitionTime":"2025-12-09T09:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.336095 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.336182 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.336204 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.336228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.336247 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:25Z","lastTransitionTime":"2025-12-09T09:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.438428 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.438474 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.438487 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.438504 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.438518 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:25Z","lastTransitionTime":"2025-12-09T09:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.542166 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.542224 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.542236 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.542255 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.542268 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:25Z","lastTransitionTime":"2025-12-09T09:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.645207 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.645278 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.645290 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.645308 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.645318 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:25Z","lastTransitionTime":"2025-12-09T09:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.748039 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.748086 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.748094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.748115 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.748125 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:25Z","lastTransitionTime":"2025-12-09T09:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.851897 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.851952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.851963 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.851983 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.851997 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:25Z","lastTransitionTime":"2025-12-09T09:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.910114 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:25 crc kubenswrapper[4824]: E1209 09:48:25.910624 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.911035 4824 scope.go:117] "RemoveContainer" containerID="352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34" Dec 09 09:48:25 crc kubenswrapper[4824]: E1209 09:48:25.911237 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vvxck_openshift-ovn-kubernetes(13620d58-6925-48c7-bbc7-f34e92922480)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" podUID="13620d58-6925-48c7-bbc7-f34e92922480" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.954528 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.954583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.954595 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.954614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:25 crc kubenswrapper[4824]: I1209 09:48:25.954627 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:25Z","lastTransitionTime":"2025-12-09T09:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.057659 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.057997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.058033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.058483 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.059652 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:26Z","lastTransitionTime":"2025-12-09T09:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.162519 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.162560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.162569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.162583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.162592 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:26Z","lastTransitionTime":"2025-12-09T09:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.265162 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.265211 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.265225 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.265244 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.265255 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:26Z","lastTransitionTime":"2025-12-09T09:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.367330 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.367361 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.367372 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.367387 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.367396 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:26Z","lastTransitionTime":"2025-12-09T09:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.469526 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.469564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.469573 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.469586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.469596 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:26Z","lastTransitionTime":"2025-12-09T09:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.572139 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.572181 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.572193 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.572210 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.572223 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:26Z","lastTransitionTime":"2025-12-09T09:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.674325 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.674362 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.674379 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.674397 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.674407 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:26Z","lastTransitionTime":"2025-12-09T09:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.777051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.777101 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.777110 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.777127 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.777744 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:26Z","lastTransitionTime":"2025-12-09T09:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.879931 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.879990 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.880001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.880021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.880033 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:26Z","lastTransitionTime":"2025-12-09T09:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.910490 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.910546 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.910652 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:26 crc kubenswrapper[4824]: E1209 09:48:26.910812 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:26 crc kubenswrapper[4824]: E1209 09:48:26.910892 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:26 crc kubenswrapper[4824]: E1209 09:48:26.910711 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.982812 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.982860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.982873 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.982891 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:26 crc kubenswrapper[4824]: I1209 09:48:26.982904 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:26Z","lastTransitionTime":"2025-12-09T09:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.086163 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.086218 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.086228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.086270 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.086293 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:27Z","lastTransitionTime":"2025-12-09T09:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.189142 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.189201 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.189212 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.189231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.189241 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:27Z","lastTransitionTime":"2025-12-09T09:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.291894 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.292670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.292732 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.292761 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.292809 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:27Z","lastTransitionTime":"2025-12-09T09:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.396181 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.396233 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.396244 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.396260 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.396271 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:27Z","lastTransitionTime":"2025-12-09T09:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.498855 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.498900 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.498910 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.498928 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.498940 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:27Z","lastTransitionTime":"2025-12-09T09:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.601499 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.601582 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.601602 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.601639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.601656 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:27Z","lastTransitionTime":"2025-12-09T09:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.704323 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.704368 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.704378 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.704396 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.704408 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:27Z","lastTransitionTime":"2025-12-09T09:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.808004 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.808062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.808076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.808099 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.808113 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:27Z","lastTransitionTime":"2025-12-09T09:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.909763 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:27 crc kubenswrapper[4824]: E1209 09:48:27.910135 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.911406 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.911473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.911489 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.911509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:27 crc kubenswrapper[4824]: I1209 09:48:27.911545 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:27Z","lastTransitionTime":"2025-12-09T09:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.014252 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.014342 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.014356 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.014377 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.014391 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:28Z","lastTransitionTime":"2025-12-09T09:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.054892 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.054944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.054955 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.054977 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.054991 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:28Z","lastTransitionTime":"2025-12-09T09:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:28 crc kubenswrapper[4824]: E1209 09:48:28.068994 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:28Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.073204 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.073262 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.073274 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.073297 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.073309 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:28Z","lastTransitionTime":"2025-12-09T09:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:28 crc kubenswrapper[4824]: E1209 09:48:28.084357 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:28Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.088531 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.088613 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.088626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.088645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.088655 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:28Z","lastTransitionTime":"2025-12-09T09:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:28 crc kubenswrapper[4824]: E1209 09:48:28.100772 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:28Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.104712 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.104760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.104774 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.104814 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.104828 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:28Z","lastTransitionTime":"2025-12-09T09:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:28 crc kubenswrapper[4824]: E1209 09:48:28.116852 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:28Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.120657 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.120693 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.120704 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.120720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.120732 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:28Z","lastTransitionTime":"2025-12-09T09:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:28 crc kubenswrapper[4824]: E1209 09:48:28.134823 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:28Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:28 crc kubenswrapper[4824]: E1209 09:48:28.134979 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.136492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.136556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.136570 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.136590 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.136603 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:28Z","lastTransitionTime":"2025-12-09T09:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.239473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.239522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.239536 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.239558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.239575 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:28Z","lastTransitionTime":"2025-12-09T09:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.342630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.342684 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.342694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.342715 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.342726 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:28Z","lastTransitionTime":"2025-12-09T09:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.445978 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.446044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.446055 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.446076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.446090 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:28Z","lastTransitionTime":"2025-12-09T09:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.550567 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.550625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.550636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.550653 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.550663 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:28Z","lastTransitionTime":"2025-12-09T09:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.653966 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.654024 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.654036 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.654054 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.654065 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:28Z","lastTransitionTime":"2025-12-09T09:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.756862 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.756931 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.756946 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.756967 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.756986 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:28Z","lastTransitionTime":"2025-12-09T09:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.861379 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.861432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.861443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.861464 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.861478 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:28Z","lastTransitionTime":"2025-12-09T09:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.910314 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.910354 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.910376 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:28 crc kubenswrapper[4824]: E1209 09:48:28.910511 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:28 crc kubenswrapper[4824]: E1209 09:48:28.910571 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:28 crc kubenswrapper[4824]: E1209 09:48:28.910723 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.964218 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.964271 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.964283 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.964304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:28 crc kubenswrapper[4824]: I1209 09:48:28.964322 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:28Z","lastTransitionTime":"2025-12-09T09:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.067440 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.067500 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.067510 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.067529 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.067540 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:29Z","lastTransitionTime":"2025-12-09T09:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.170339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.170437 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.170454 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.170478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.170496 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:29Z","lastTransitionTime":"2025-12-09T09:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.272985 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.273047 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.273058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.273077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.273092 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:29Z","lastTransitionTime":"2025-12-09T09:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.376340 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.376390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.376401 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.376421 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.376434 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:29Z","lastTransitionTime":"2025-12-09T09:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.478647 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.478694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.478706 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.478735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.478746 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:29Z","lastTransitionTime":"2025-12-09T09:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.581320 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.581369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.581383 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.581398 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.581410 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:29Z","lastTransitionTime":"2025-12-09T09:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.684385 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.684429 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.684446 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.684464 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.684476 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:29Z","lastTransitionTime":"2025-12-09T09:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.787635 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.787700 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.787724 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.787756 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.787812 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:29Z","lastTransitionTime":"2025-12-09T09:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.891004 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.891044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.891058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.891076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.891086 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:29Z","lastTransitionTime":"2025-12-09T09:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.909867 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:29 crc kubenswrapper[4824]: E1209 09:48:29.910037 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.993937 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.993999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.994008 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.994042 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:29 crc kubenswrapper[4824]: I1209 09:48:29.994054 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:29Z","lastTransitionTime":"2025-12-09T09:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.096581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.096622 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.096633 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.096651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.096662 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:30Z","lastTransitionTime":"2025-12-09T09:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.199120 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.199170 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.199182 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.199201 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.199213 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:30Z","lastTransitionTime":"2025-12-09T09:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.301409 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.301458 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.301472 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.301492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.301503 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:30Z","lastTransitionTime":"2025-12-09T09:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.403655 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.403736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.403757 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.403807 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.403819 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:30Z","lastTransitionTime":"2025-12-09T09:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.488557 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs\") pod \"network-metrics-daemon-5tn52\" (UID: \"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\") " pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:30 crc kubenswrapper[4824]: E1209 09:48:30.488721 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 09:48:30 crc kubenswrapper[4824]: E1209 09:48:30.488805 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs podName:fcbce9ea-3d55-4e09-b5bc-d5bb8c472219 nodeName:}" failed. No retries permitted until 2025-12-09 09:49:02.488763497 +0000 UTC m=+98.823268164 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs") pod "network-metrics-daemon-5tn52" (UID: "fcbce9ea-3d55-4e09-b5bc-d5bb8c472219") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.506573 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.506613 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.506643 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.506664 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.506674 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:30Z","lastTransitionTime":"2025-12-09T09:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.609250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.609303 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.609313 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.609329 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.609342 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:30Z","lastTransitionTime":"2025-12-09T09:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.712431 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.712508 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.712518 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.712539 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.712552 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:30Z","lastTransitionTime":"2025-12-09T09:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.815377 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.815427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.815441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.815461 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.815475 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:30Z","lastTransitionTime":"2025-12-09T09:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.910364 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.910457 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.910457 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:30 crc kubenswrapper[4824]: E1209 09:48:30.910600 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:30 crc kubenswrapper[4824]: E1209 09:48:30.910654 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:30 crc kubenswrapper[4824]: E1209 09:48:30.910738 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.918589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.918654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.918666 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.918686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:30 crc kubenswrapper[4824]: I1209 09:48:30.918698 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:30Z","lastTransitionTime":"2025-12-09T09:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.021775 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.021899 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.021919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.021945 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.021965 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:31Z","lastTransitionTime":"2025-12-09T09:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.125808 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.126209 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.126347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.126526 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.126675 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:31Z","lastTransitionTime":"2025-12-09T09:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.230126 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.230172 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.230183 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.230202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.230218 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:31Z","lastTransitionTime":"2025-12-09T09:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.333598 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.333683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.333708 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.333740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.333769 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:31Z","lastTransitionTime":"2025-12-09T09:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.436626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.436697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.436713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.436741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.436761 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:31Z","lastTransitionTime":"2025-12-09T09:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.539752 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.539828 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.539838 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.539858 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.539871 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:31Z","lastTransitionTime":"2025-12-09T09:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.643024 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.643075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.643086 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.643106 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.643120 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:31Z","lastTransitionTime":"2025-12-09T09:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.746354 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.746397 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.746408 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.746426 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.746436 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:31Z","lastTransitionTime":"2025-12-09T09:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.850090 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.850161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.850178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.850214 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.850230 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:31Z","lastTransitionTime":"2025-12-09T09:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.909854 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:31 crc kubenswrapper[4824]: E1209 09:48:31.910041 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.952677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.952736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.952747 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.952813 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:31 crc kubenswrapper[4824]: I1209 09:48:31.952889 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:31Z","lastTransitionTime":"2025-12-09T09:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.055576 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.055631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.055644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.055663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.055675 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:32Z","lastTransitionTime":"2025-12-09T09:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.158955 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.159021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.159034 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.159060 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.159072 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:32Z","lastTransitionTime":"2025-12-09T09:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.262062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.262129 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.262141 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.262160 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.262174 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:32Z","lastTransitionTime":"2025-12-09T09:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.330270 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-pvksg_74aa3cd7-0abc-4fc9-9697-e23b693a7611/kube-multus/0.log" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.330339 4824 generic.go:334] "Generic (PLEG): container finished" podID="74aa3cd7-0abc-4fc9-9697-e23b693a7611" containerID="cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de" exitCode=1 Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.330380 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pvksg" event={"ID":"74aa3cd7-0abc-4fc9-9697-e23b693a7611","Type":"ContainerDied","Data":"cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de"} Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.330878 4824 scope.go:117] "RemoveContainer" containerID="cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.351331 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2dd916-62ee-4dca-964d-fb859475a9d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1e6cfeb997cafb7ec7f6d829e1ff2e2d95e4b343c422194a5290d222ea53a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2add2743e390cf2b91c9b020de38fc6df392fee9e5aa1d59bd1ca9471d9d8d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17601a452dc4369a533806384d58f733e396e255f5741b38004cae98a44cfe55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eb98ce1943d5692f7e6d51befb0c05115ee403be72134f40f651a80bf0a378b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8eb98ce1943d5692f7e6d51befb0c05115ee403be72134f40f651a80bf0a378b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:32Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.366574 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.366623 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.366639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.366661 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.366678 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:32Z","lastTransitionTime":"2025-12-09T09:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.370145 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:32Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.385346 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:32Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.401503 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:48:31Z\\\",\\\"message\\\":\\\"2025-12-09T09:47:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c3558da8-38fa-42ec-b5cd-4f174f21b210\\\\n2025-12-09T09:47:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c3558da8-38fa-42ec-b5cd-4f174f21b210 to /host/opt/cni/bin/\\\\n2025-12-09T09:47:46Z [verbose] multus-daemon started\\\\n2025-12-09T09:47:46Z [verbose] Readiness Indicator file check\\\\n2025-12-09T09:48:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:32Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.415634 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:32Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.431378 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:32Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.451185 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:48:11Z\\\",\\\"message\\\":\\\"ncer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI1209 09:48:11.888995 6487 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 2.013563ms\\\\nI1209 09:48:11.889004 6487 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1209 09:48:11.889013 6487 services_controller.go:356] Processing sync for service openshift-operator-lifecycle-manager/packageserver-service for network=default\\\\nI1209 09:48:11.889016 6487 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1209 09:48:11.889010 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": fail\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:48:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vvxck_openshift-ovn-kubernetes(13620d58-6925-48c7-bbc7-f34e92922480)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:32Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.462718 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daad2ae578b4f337ceb3298b012b0262903cebe4e9e9584513cf2f31fbcec3c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37860b2742b9486e06919b4bb4e7e4631ff5905eea837ef17c1d2f8d573da18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gcrqx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:32Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.468603 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.468637 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.468648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.468665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.468677 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:32Z","lastTransitionTime":"2025-12-09T09:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.474404 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5tn52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5tn52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:32Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.494571 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:32Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.509120 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:32Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.523442 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:32Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.537081 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:32Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.548942 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:32Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.562470 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:32Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.571217 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.571259 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.571297 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.571329 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.571344 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:32Z","lastTransitionTime":"2025-12-09T09:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.577584 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:32Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.592176 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:32Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.604724 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:32Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.680516 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.680572 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.680604 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.680628 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.680642 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:32Z","lastTransitionTime":"2025-12-09T09:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.789392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.789451 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.789461 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.789475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.789485 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:32Z","lastTransitionTime":"2025-12-09T09:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.892612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.892658 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.892668 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.892687 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.892698 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:32Z","lastTransitionTime":"2025-12-09T09:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.910025 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:32 crc kubenswrapper[4824]: E1209 09:48:32.910214 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.910517 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:32 crc kubenswrapper[4824]: E1209 09:48:32.910626 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.910884 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:32 crc kubenswrapper[4824]: E1209 09:48:32.910989 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.995875 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.995914 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.995922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.995938 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:32 crc kubenswrapper[4824]: I1209 09:48:32.995948 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:32Z","lastTransitionTime":"2025-12-09T09:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.098684 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.098731 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.098744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.098762 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.098775 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:33Z","lastTransitionTime":"2025-12-09T09:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.201676 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.201735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.201745 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.201764 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.201775 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:33Z","lastTransitionTime":"2025-12-09T09:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.304749 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.304837 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.304848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.304870 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.304886 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:33Z","lastTransitionTime":"2025-12-09T09:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.335743 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-pvksg_74aa3cd7-0abc-4fc9-9697-e23b693a7611/kube-multus/0.log" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.335823 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pvksg" event={"ID":"74aa3cd7-0abc-4fc9-9697-e23b693a7611","Type":"ContainerStarted","Data":"c0ee66e89363676597efd13d02245ce8c9a9af288c04f2ed9b7ae8bed31da259"} Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.359305 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:33Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.373500 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:33Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.387175 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:33Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.400758 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:33Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.408047 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.408086 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.408096 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.408114 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.408126 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:33Z","lastTransitionTime":"2025-12-09T09:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.412925 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:33Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.426653 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:33Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.440646 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:33Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.451726 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:33Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.466399 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:33Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.481171 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:33Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.495267 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:33Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.510603 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0ee66e89363676597efd13d02245ce8c9a9af288c04f2ed9b7ae8bed31da259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:48:31Z\\\",\\\"message\\\":\\\"2025-12-09T09:47:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c3558da8-38fa-42ec-b5cd-4f174f21b210\\\\n2025-12-09T09:47:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c3558da8-38fa-42ec-b5cd-4f174f21b210 to /host/opt/cni/bin/\\\\n2025-12-09T09:47:46Z [verbose] multus-daemon started\\\\n2025-12-09T09:47:46Z [verbose] Readiness Indicator file check\\\\n2025-12-09T09:48:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:33Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.511585 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.511634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.511650 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.511674 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.511685 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:33Z","lastTransitionTime":"2025-12-09T09:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.525089 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2dd916-62ee-4dca-964d-fb859475a9d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1e6cfeb997cafb7ec7f6d829e1ff2e2d95e4b343c422194a5290d222ea53a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2add2743e390cf2b91c9b020de38fc6df392fee9e5aa1d59bd1ca9471d9d8d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17601a452dc4369a533806384d58f733e396e255f5741b38004cae98a44cfe55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eb98ce1943d5692f7e6d51befb0c05115ee403be72134f40f651a80bf0a378b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8eb98ce1943d5692f7e6d51befb0c05115ee403be72134f40f651a80bf0a378b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:33Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.545266 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:33Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.568043 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:48:11Z\\\",\\\"message\\\":\\\"ncer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI1209 09:48:11.888995 6487 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 2.013563ms\\\\nI1209 09:48:11.889004 6487 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1209 09:48:11.889013 6487 services_controller.go:356] Processing sync for service openshift-operator-lifecycle-manager/packageserver-service for network=default\\\\nI1209 09:48:11.889016 6487 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1209 09:48:11.889010 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": fail\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:48:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vvxck_openshift-ovn-kubernetes(13620d58-6925-48c7-bbc7-f34e92922480)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:33Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.580989 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daad2ae578b4f337ceb3298b012b0262903cebe4e9e9584513cf2f31fbcec3c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37860b2742b9486e06919b4bb4e7e4631ff5905eea837ef17c1d2f8d573da18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gcrqx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:33Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.594980 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5tn52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5tn52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:33Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.609706 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:33Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.615029 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.615078 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.615107 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.615132 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.615145 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:33Z","lastTransitionTime":"2025-12-09T09:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.720176 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.720240 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.720255 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.720282 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.720295 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:33Z","lastTransitionTime":"2025-12-09T09:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.823537 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.823577 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.823589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.823607 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.823619 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:33Z","lastTransitionTime":"2025-12-09T09:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.911057 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:33 crc kubenswrapper[4824]: E1209 09:48:33.911217 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.926121 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.926152 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.926164 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.926180 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.926191 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:33Z","lastTransitionTime":"2025-12-09T09:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.944074 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:33Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.961108 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:33Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.978579 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:33Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:33 crc kubenswrapper[4824]: I1209 09:48:33.995400 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:33Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.007635 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:34Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.020338 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:34Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.027772 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.027827 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.027839 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.027854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.027864 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:34Z","lastTransitionTime":"2025-12-09T09:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.034725 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:34Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.049434 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:34Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.060653 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:34Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.072687 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2dd916-62ee-4dca-964d-fb859475a9d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1e6cfeb997cafb7ec7f6d829e1ff2e2d95e4b343c422194a5290d222ea53a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2add2743e390cf2b91c9b020de38fc6df392fee9e5aa1d59bd1ca9471d9d8d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17601a452dc4369a533806384d58f733e396e255f5741b38004cae98a44cfe55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eb98ce1943d5692f7e6d51befb0c05115ee403be72134f40f651a80bf0a378b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8eb98ce1943d5692f7e6d51befb0c05115ee403be72134f40f651a80bf0a378b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:34Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.085507 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:34Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.100054 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:34Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.114761 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0ee66e89363676597efd13d02245ce8c9a9af288c04f2ed9b7ae8bed31da259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:48:31Z\\\",\\\"message\\\":\\\"2025-12-09T09:47:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c3558da8-38fa-42ec-b5cd-4f174f21b210\\\\n2025-12-09T09:47:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c3558da8-38fa-42ec-b5cd-4f174f21b210 to /host/opt/cni/bin/\\\\n2025-12-09T09:47:46Z [verbose] multus-daemon started\\\\n2025-12-09T09:47:46Z [verbose] Readiness Indicator file check\\\\n2025-12-09T09:48:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:34Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.128747 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:34Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.129588 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.129619 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.129629 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.129645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.129659 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:34Z","lastTransitionTime":"2025-12-09T09:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.145398 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:34Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.169808 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:48:11Z\\\",\\\"message\\\":\\\"ncer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI1209 09:48:11.888995 6487 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 2.013563ms\\\\nI1209 09:48:11.889004 6487 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1209 09:48:11.889013 6487 services_controller.go:356] Processing sync for service openshift-operator-lifecycle-manager/packageserver-service for network=default\\\\nI1209 09:48:11.889016 6487 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1209 09:48:11.889010 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": fail\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:48:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vvxck_openshift-ovn-kubernetes(13620d58-6925-48c7-bbc7-f34e92922480)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:34Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.180936 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daad2ae578b4f337ceb3298b012b0262903cebe4e9e9584513cf2f31fbcec3c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37860b2742b9486e06919b4bb4e7e4631ff5905eea837ef17c1d2f8d573da18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gcrqx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:34Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.191651 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5tn52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5tn52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:34Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.231684 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.231721 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.231732 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.231747 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.231756 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:34Z","lastTransitionTime":"2025-12-09T09:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.335335 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.335402 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.335435 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.335455 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.335467 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:34Z","lastTransitionTime":"2025-12-09T09:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.437912 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.437974 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.437990 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.438012 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.438031 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:34Z","lastTransitionTime":"2025-12-09T09:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.540551 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.540609 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.540622 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.540641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.540657 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:34Z","lastTransitionTime":"2025-12-09T09:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.642644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.642682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.642694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.642712 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.642725 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:34Z","lastTransitionTime":"2025-12-09T09:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.746180 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.746239 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.746250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.746270 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.746284 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:34Z","lastTransitionTime":"2025-12-09T09:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.849486 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.849553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.849565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.849587 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.849602 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:34Z","lastTransitionTime":"2025-12-09T09:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.910198 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.910198 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.910330 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:34 crc kubenswrapper[4824]: E1209 09:48:34.910588 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:34 crc kubenswrapper[4824]: E1209 09:48:34.910678 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:34 crc kubenswrapper[4824]: E1209 09:48:34.910803 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.952274 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.952316 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.952325 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.952339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:34 crc kubenswrapper[4824]: I1209 09:48:34.952348 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:34Z","lastTransitionTime":"2025-12-09T09:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.055564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.055606 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.055617 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.055632 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.055641 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:35Z","lastTransitionTime":"2025-12-09T09:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.158191 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.158248 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.158261 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.158281 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.158298 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:35Z","lastTransitionTime":"2025-12-09T09:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.261376 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.261422 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.261431 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.261447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.261458 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:35Z","lastTransitionTime":"2025-12-09T09:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.363746 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.363824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.363837 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.363855 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.363868 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:35Z","lastTransitionTime":"2025-12-09T09:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.466738 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.466797 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.466810 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.466829 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.466840 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:35Z","lastTransitionTime":"2025-12-09T09:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.569583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.569630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.569639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.569655 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.569668 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:35Z","lastTransitionTime":"2025-12-09T09:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.672284 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.672331 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.672342 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.672360 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.672373 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:35Z","lastTransitionTime":"2025-12-09T09:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.774722 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.774801 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.774814 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.774831 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.774844 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:35Z","lastTransitionTime":"2025-12-09T09:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.877000 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.877405 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.877415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.877433 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.877444 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:35Z","lastTransitionTime":"2025-12-09T09:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.909422 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:35 crc kubenswrapper[4824]: E1209 09:48:35.909676 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.979709 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.979748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.979757 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.979771 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:35 crc kubenswrapper[4824]: I1209 09:48:35.979798 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:35Z","lastTransitionTime":"2025-12-09T09:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.082614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.082658 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.082667 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.082683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.082695 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:36Z","lastTransitionTime":"2025-12-09T09:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.185582 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.185663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.185682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.185707 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.185724 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:36Z","lastTransitionTime":"2025-12-09T09:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.288452 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.288502 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.288514 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.288530 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.288546 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:36Z","lastTransitionTime":"2025-12-09T09:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.391443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.391550 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.391579 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.391612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.391640 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:36Z","lastTransitionTime":"2025-12-09T09:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.494554 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.494605 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.494617 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.494643 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.494657 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:36Z","lastTransitionTime":"2025-12-09T09:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.598078 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.598163 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.598177 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.598198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.598212 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:36Z","lastTransitionTime":"2025-12-09T09:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.701432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.701471 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.701480 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.701498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.701515 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:36Z","lastTransitionTime":"2025-12-09T09:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.803742 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.803798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.803812 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.803827 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.803839 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:36Z","lastTransitionTime":"2025-12-09T09:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.906635 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.906698 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.906712 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.906738 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.906753 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:36Z","lastTransitionTime":"2025-12-09T09:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.909878 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.909894 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:36 crc kubenswrapper[4824]: I1209 09:48:36.910034 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:36 crc kubenswrapper[4824]: E1209 09:48:36.910221 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:36 crc kubenswrapper[4824]: E1209 09:48:36.910508 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:36 crc kubenswrapper[4824]: E1209 09:48:36.910638 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.009388 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.009422 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.009430 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.009444 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.009454 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:37Z","lastTransitionTime":"2025-12-09T09:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.111915 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.112009 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.112043 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.112077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.112098 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:37Z","lastTransitionTime":"2025-12-09T09:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.214388 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.214437 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.214449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.214470 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.214483 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:37Z","lastTransitionTime":"2025-12-09T09:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.317126 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.317173 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.317182 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.317197 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.317207 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:37Z","lastTransitionTime":"2025-12-09T09:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.419355 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.419401 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.419412 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.419427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.419444 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:37Z","lastTransitionTime":"2025-12-09T09:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.526490 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.526557 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.526568 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.526589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.526598 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:37Z","lastTransitionTime":"2025-12-09T09:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.629836 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.629919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.629937 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.630406 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.630463 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:37Z","lastTransitionTime":"2025-12-09T09:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.733704 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.733756 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.733768 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.733798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.733810 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:37Z","lastTransitionTime":"2025-12-09T09:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.836245 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.836317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.836331 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.836347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.836358 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:37Z","lastTransitionTime":"2025-12-09T09:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.911468 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:37 crc kubenswrapper[4824]: E1209 09:48:37.911664 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.938997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.939034 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.939046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.939062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:37 crc kubenswrapper[4824]: I1209 09:48:37.939072 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:37Z","lastTransitionTime":"2025-12-09T09:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.043686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.043740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.043753 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.043773 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.043812 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:38Z","lastTransitionTime":"2025-12-09T09:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.146491 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.146532 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.146543 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.146561 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.146577 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:38Z","lastTransitionTime":"2025-12-09T09:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.214960 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.215021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.215037 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.215063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.215079 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:38Z","lastTransitionTime":"2025-12-09T09:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:38 crc kubenswrapper[4824]: E1209 09:48:38.228948 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:38Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.233655 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.233696 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.233705 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.233721 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.233736 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:38Z","lastTransitionTime":"2025-12-09T09:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:38 crc kubenswrapper[4824]: E1209 09:48:38.248242 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:38Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.253111 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.253154 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.253164 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.253186 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.253205 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:38Z","lastTransitionTime":"2025-12-09T09:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:38 crc kubenswrapper[4824]: E1209 09:48:38.266084 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:38Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.269744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.269844 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.269868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.269898 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.269921 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:38Z","lastTransitionTime":"2025-12-09T09:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:38 crc kubenswrapper[4824]: E1209 09:48:38.283538 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:38Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.288122 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.288193 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.288244 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.288272 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.288292 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:38Z","lastTransitionTime":"2025-12-09T09:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:38 crc kubenswrapper[4824]: E1209 09:48:38.302654 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"047c928d-3ca9-4a45-942a-c245de14f559\\\",\\\"systemUUID\\\":\\\"12ef6ade-9e0b-4b45-ad21-e9edcb9cfb09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:38Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:38 crc kubenswrapper[4824]: E1209 09:48:38.302857 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.304419 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.304463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.304480 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.304504 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.304517 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:38Z","lastTransitionTime":"2025-12-09T09:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.408032 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.408083 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.408100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.408129 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.408142 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:38Z","lastTransitionTime":"2025-12-09T09:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.512237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.512314 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.512370 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.512397 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.512416 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:38Z","lastTransitionTime":"2025-12-09T09:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.615845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.615899 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.615913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.615933 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.615947 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:38Z","lastTransitionTime":"2025-12-09T09:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.721309 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.721389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.721415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.721445 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.721468 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:38Z","lastTransitionTime":"2025-12-09T09:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.824722 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.824859 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.824897 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.824932 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.824956 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:38Z","lastTransitionTime":"2025-12-09T09:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.910092 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:38 crc kubenswrapper[4824]: E1209 09:48:38.910908 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.910183 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:38 crc kubenswrapper[4824]: E1209 09:48:38.910993 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.910104 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:38 crc kubenswrapper[4824]: E1209 09:48:38.911067 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.927306 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.927358 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.927371 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.927389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:38 crc kubenswrapper[4824]: I1209 09:48:38.927401 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:38Z","lastTransitionTime":"2025-12-09T09:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.030613 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.030710 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.030738 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.030770 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.030881 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:39Z","lastTransitionTime":"2025-12-09T09:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.134079 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.134136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.134152 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.134176 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.134193 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:39Z","lastTransitionTime":"2025-12-09T09:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.237581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.237641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.237652 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.237683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.237700 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:39Z","lastTransitionTime":"2025-12-09T09:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.340645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.340687 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.340701 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.340722 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.340735 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:39Z","lastTransitionTime":"2025-12-09T09:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.444133 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.444197 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.444208 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.444225 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.444239 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:39Z","lastTransitionTime":"2025-12-09T09:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.547302 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.547389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.547413 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.547451 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.547473 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:39Z","lastTransitionTime":"2025-12-09T09:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.651305 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.651375 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.651392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.651417 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.651434 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:39Z","lastTransitionTime":"2025-12-09T09:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.754552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.754600 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.754609 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.754624 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.754635 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:39Z","lastTransitionTime":"2025-12-09T09:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.857455 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.857511 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.857524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.857547 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.857562 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:39Z","lastTransitionTime":"2025-12-09T09:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.910407 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:39 crc kubenswrapper[4824]: E1209 09:48:39.910608 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.960502 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.960541 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.960551 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.960567 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:39 crc kubenswrapper[4824]: I1209 09:48:39.960577 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:39Z","lastTransitionTime":"2025-12-09T09:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.063474 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.063514 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.063524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.063537 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.063546 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:40Z","lastTransitionTime":"2025-12-09T09:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.166613 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.166662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.166672 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.166694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.166707 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:40Z","lastTransitionTime":"2025-12-09T09:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.268872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.268919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.268929 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.268944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.268954 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:40Z","lastTransitionTime":"2025-12-09T09:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.371252 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.371298 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.371319 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.371340 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.371354 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:40Z","lastTransitionTime":"2025-12-09T09:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.473321 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.473415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.473428 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.473449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.473468 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:40Z","lastTransitionTime":"2025-12-09T09:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.575978 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.576063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.576087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.576122 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.576151 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:40Z","lastTransitionTime":"2025-12-09T09:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.678603 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.678645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.678657 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.678674 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.678686 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:40Z","lastTransitionTime":"2025-12-09T09:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.781096 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.781137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.781147 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.781164 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.781175 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:40Z","lastTransitionTime":"2025-12-09T09:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.883526 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.883577 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.883589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.883608 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.883622 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:40Z","lastTransitionTime":"2025-12-09T09:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.910107 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.910218 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:40 crc kubenswrapper[4824]: E1209 09:48:40.910352 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.910400 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:40 crc kubenswrapper[4824]: E1209 09:48:40.910555 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:40 crc kubenswrapper[4824]: E1209 09:48:40.910758 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.911625 4824 scope.go:117] "RemoveContainer" containerID="352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.986618 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.986648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.986677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.986691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:40 crc kubenswrapper[4824]: I1209 09:48:40.986701 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:40Z","lastTransitionTime":"2025-12-09T09:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.089483 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.089534 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.089546 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.089568 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.089604 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:41Z","lastTransitionTime":"2025-12-09T09:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.193236 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.193304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.193316 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.193335 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.193346 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:41Z","lastTransitionTime":"2025-12-09T09:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.296013 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.296058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.296070 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.296089 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.296099 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:41Z","lastTransitionTime":"2025-12-09T09:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.369980 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vvxck_13620d58-6925-48c7-bbc7-f34e92922480/ovnkube-controller/2.log" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.372922 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerStarted","Data":"55659b94cb673db2b20bd35a15232443b44882962e7731116b74c418ea5d251e"} Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.374092 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.387340 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:41Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.398595 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.398636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.398645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.398666 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.398682 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:41Z","lastTransitionTime":"2025-12-09T09:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.398773 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:41Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.424811 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:41Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.484880 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:41Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.501328 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.501372 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.501385 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.501400 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.501410 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:41Z","lastTransitionTime":"2025-12-09T09:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.503200 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:41Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.519482 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:41Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.533523 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:41Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.547269 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:41Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.557488 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:41Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.567349 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2dd916-62ee-4dca-964d-fb859475a9d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1e6cfeb997cafb7ec7f6d829e1ff2e2d95e4b343c422194a5290d222ea53a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2add2743e390cf2b91c9b020de38fc6df392fee9e5aa1d59bd1ca9471d9d8d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17601a452dc4369a533806384d58f733e396e255f5741b38004cae98a44cfe55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eb98ce1943d5692f7e6d51befb0c05115ee403be72134f40f651a80bf0a378b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8eb98ce1943d5692f7e6d51befb0c05115ee403be72134f40f651a80bf0a378b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:41Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.578556 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:41Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.589254 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:41Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.603710 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.603790 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.603801 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.603824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.603846 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:41Z","lastTransitionTime":"2025-12-09T09:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.603893 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0ee66e89363676597efd13d02245ce8c9a9af288c04f2ed9b7ae8bed31da259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:48:31Z\\\",\\\"message\\\":\\\"2025-12-09T09:47:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c3558da8-38fa-42ec-b5cd-4f174f21b210\\\\n2025-12-09T09:47:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c3558da8-38fa-42ec-b5cd-4f174f21b210 to /host/opt/cni/bin/\\\\n2025-12-09T09:47:46Z [verbose] multus-daemon started\\\\n2025-12-09T09:47:46Z [verbose] Readiness Indicator file check\\\\n2025-12-09T09:48:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:41Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.615005 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5tn52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5tn52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:41Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.627314 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:41Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.642319 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:41Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.663355 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55659b94cb673db2b20bd35a15232443b44882962e7731116b74c418ea5d251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:48:11Z\\\",\\\"message\\\":\\\"ncer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI1209 09:48:11.888995 6487 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 2.013563ms\\\\nI1209 09:48:11.889004 6487 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1209 09:48:11.889013 6487 services_controller.go:356] Processing sync for service openshift-operator-lifecycle-manager/packageserver-service for network=default\\\\nI1209 09:48:11.889016 6487 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1209 09:48:11.889010 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": fail\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:48:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:48:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:41Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.676775 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daad2ae578b4f337ceb3298b012b0262903cebe4e9e9584513cf2f31fbcec3c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37860b2742b9486e06919b4bb4e7e4631ff5905eea837ef17c1d2f8d573da18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gcrqx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:41Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.706774 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.706833 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.706845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.706860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.706871 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:41Z","lastTransitionTime":"2025-12-09T09:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.809817 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.809893 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.809908 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.809936 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.809952 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:41Z","lastTransitionTime":"2025-12-09T09:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.909845 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:41 crc kubenswrapper[4824]: E1209 09:48:41.910048 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.912428 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.912490 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.912499 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.912518 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:41 crc kubenswrapper[4824]: I1209 09:48:41.912529 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:41Z","lastTransitionTime":"2025-12-09T09:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.015890 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.015941 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.015951 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.015971 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.015982 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:42Z","lastTransitionTime":"2025-12-09T09:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.118708 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.118760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.118775 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.118819 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.118834 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:42Z","lastTransitionTime":"2025-12-09T09:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.223371 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.223472 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.223503 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.223542 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.223580 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:42Z","lastTransitionTime":"2025-12-09T09:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.325963 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.326007 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.326017 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.326033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.326043 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:42Z","lastTransitionTime":"2025-12-09T09:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.377315 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vvxck_13620d58-6925-48c7-bbc7-f34e92922480/ovnkube-controller/3.log" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.378012 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vvxck_13620d58-6925-48c7-bbc7-f34e92922480/ovnkube-controller/2.log" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.380402 4824 generic.go:334] "Generic (PLEG): container finished" podID="13620d58-6925-48c7-bbc7-f34e92922480" containerID="55659b94cb673db2b20bd35a15232443b44882962e7731116b74c418ea5d251e" exitCode=1 Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.380438 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerDied","Data":"55659b94cb673db2b20bd35a15232443b44882962e7731116b74c418ea5d251e"} Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.380476 4824 scope.go:117] "RemoveContainer" containerID="352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.381342 4824 scope.go:117] "RemoveContainer" containerID="55659b94cb673db2b20bd35a15232443b44882962e7731116b74c418ea5d251e" Dec 09 09:48:42 crc kubenswrapper[4824]: E1209 09:48:42.381556 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vvxck_openshift-ovn-kubernetes(13620d58-6925-48c7-bbc7-f34e92922480)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" podUID="13620d58-6925-48c7-bbc7-f34e92922480" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.394666 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:42Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.406941 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:42Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.417498 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:42Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.428841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.428909 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.428919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.428934 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.428945 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:42Z","lastTransitionTime":"2025-12-09T09:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.436279 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:42Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.452362 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:42Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.466841 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:42Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.481486 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:42Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.496027 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:42Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.510604 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:42Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.524967 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2dd916-62ee-4dca-964d-fb859475a9d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1e6cfeb997cafb7ec7f6d829e1ff2e2d95e4b343c422194a5290d222ea53a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2add2743e390cf2b91c9b020de38fc6df392fee9e5aa1d59bd1ca9471d9d8d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17601a452dc4369a533806384d58f733e396e255f5741b38004cae98a44cfe55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eb98ce1943d5692f7e6d51befb0c05115ee403be72134f40f651a80bf0a378b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8eb98ce1943d5692f7e6d51befb0c05115ee403be72134f40f651a80bf0a378b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:42Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.532022 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.532077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.532090 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.532110 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.532126 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:42Z","lastTransitionTime":"2025-12-09T09:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.539525 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:42Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.556113 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:42Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.573144 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0ee66e89363676597efd13d02245ce8c9a9af288c04f2ed9b7ae8bed31da259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:48:31Z\\\",\\\"message\\\":\\\"2025-12-09T09:47:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c3558da8-38fa-42ec-b5cd-4f174f21b210\\\\n2025-12-09T09:47:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c3558da8-38fa-42ec-b5cd-4f174f21b210 to /host/opt/cni/bin/\\\\n2025-12-09T09:47:46Z [verbose] multus-daemon started\\\\n2025-12-09T09:47:46Z [verbose] Readiness Indicator file check\\\\n2025-12-09T09:48:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:42Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.586575 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5tn52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5tn52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:42Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.601453 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:42Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.617071 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:42Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.635015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.635081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.635096 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.635118 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.635134 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:42Z","lastTransitionTime":"2025-12-09T09:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.638754 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55659b94cb673db2b20bd35a15232443b44882962e7731116b74c418ea5d251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://352c9497a04079eb4395d80ef7667a70c7ca65130e807bda41a910938fcc5b34\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:48:11Z\\\",\\\"message\\\":\\\"ncer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI1209 09:48:11.888995 6487 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 2.013563ms\\\\nI1209 09:48:11.889004 6487 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1209 09:48:11.889013 6487 services_controller.go:356] Processing sync for service openshift-operator-lifecycle-manager/packageserver-service for network=default\\\\nI1209 09:48:11.889016 6487 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1209 09:48:11.889010 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": fail\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:48:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55659b94cb673db2b20bd35a15232443b44882962e7731116b74c418ea5d251e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:48:42Z\\\",\\\"message\\\":\\\"ceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1209 09:48:41.874805 6920 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx in node crc\\\\nI1209 09:48:41.874813 6920 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx after 0 failed attempt(s)\\\\nI1209 09:48:41.874821 6920 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx\\\\nI1209 09:48:41.874844 6920 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1209 09:48:41.874869 6920 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}\\\\nI1209 09:48:41.874881 6920 services_controller.go:360] Finished syncing service olm-operator-metrics on namespace openshift-operator-lifecycle-manager for network=default : 7.206478ms\\\\nI1209 09:48:41.874892 6920 services_controller.go:356] Processing sync for service openshift-dns/dns-default for network=default\\\\nF1209 09:48:41.874931 6920 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:48:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:42Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.653064 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daad2ae578b4f337ceb3298b012b0262903cebe4e9e9584513cf2f31fbcec3c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37860b2742b9486e06919b4bb4e7e4631ff5905eea837ef17c1d2f8d573da18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gcrqx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:42Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.737847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.737887 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.737896 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.737911 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.737923 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:42Z","lastTransitionTime":"2025-12-09T09:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.840614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.840657 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.840670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.840692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.840710 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:42Z","lastTransitionTime":"2025-12-09T09:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.910120 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.910185 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.910143 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:42 crc kubenswrapper[4824]: E1209 09:48:42.910373 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:42 crc kubenswrapper[4824]: E1209 09:48:42.910566 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:42 crc kubenswrapper[4824]: E1209 09:48:42.910851 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.944106 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.944151 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.944161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.944179 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:42 crc kubenswrapper[4824]: I1209 09:48:42.944190 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:42Z","lastTransitionTime":"2025-12-09T09:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.046136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.046177 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.046189 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.046206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.046218 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:43Z","lastTransitionTime":"2025-12-09T09:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.149549 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.149619 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.149639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.149664 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.149680 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:43Z","lastTransitionTime":"2025-12-09T09:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.252287 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.252346 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.252362 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.252382 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.252395 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:43Z","lastTransitionTime":"2025-12-09T09:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.354845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.354884 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.354930 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.354952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.354964 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:43Z","lastTransitionTime":"2025-12-09T09:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.385984 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vvxck_13620d58-6925-48c7-bbc7-f34e92922480/ovnkube-controller/3.log" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.389410 4824 scope.go:117] "RemoveContainer" containerID="55659b94cb673db2b20bd35a15232443b44882962e7731116b74c418ea5d251e" Dec 09 09:48:43 crc kubenswrapper[4824]: E1209 09:48:43.389573 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vvxck_openshift-ovn-kubernetes(13620d58-6925-48c7-bbc7-f34e92922480)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" podUID="13620d58-6925-48c7-bbc7-f34e92922480" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.402887 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.414324 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.424392 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.445742 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.459340 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.459415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.459436 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.459465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.459486 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:43Z","lastTransitionTime":"2025-12-09T09:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.468719 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.482216 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.494916 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.506190 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.515535 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.526326 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2dd916-62ee-4dca-964d-fb859475a9d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1e6cfeb997cafb7ec7f6d829e1ff2e2d95e4b343c422194a5290d222ea53a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2add2743e390cf2b91c9b020de38fc6df392fee9e5aa1d59bd1ca9471d9d8d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17601a452dc4369a533806384d58f733e396e255f5741b38004cae98a44cfe55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eb98ce1943d5692f7e6d51befb0c05115ee403be72134f40f651a80bf0a378b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8eb98ce1943d5692f7e6d51befb0c05115ee403be72134f40f651a80bf0a378b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.539983 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.555063 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.562843 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.562880 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.562894 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.562914 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.562927 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:43Z","lastTransitionTime":"2025-12-09T09:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.566362 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0ee66e89363676597efd13d02245ce8c9a9af288c04f2ed9b7ae8bed31da259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:48:31Z\\\",\\\"message\\\":\\\"2025-12-09T09:47:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c3558da8-38fa-42ec-b5cd-4f174f21b210\\\\n2025-12-09T09:47:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c3558da8-38fa-42ec-b5cd-4f174f21b210 to /host/opt/cni/bin/\\\\n2025-12-09T09:47:46Z [verbose] multus-daemon started\\\\n2025-12-09T09:47:46Z [verbose] Readiness Indicator file check\\\\n2025-12-09T09:48:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.574537 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5tn52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5tn52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.583715 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.596309 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.633709 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55659b94cb673db2b20bd35a15232443b44882962e7731116b74c418ea5d251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55659b94cb673db2b20bd35a15232443b44882962e7731116b74c418ea5d251e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:48:42Z\\\",\\\"message\\\":\\\"ceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1209 09:48:41.874805 6920 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx in node crc\\\\nI1209 09:48:41.874813 6920 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx after 0 failed attempt(s)\\\\nI1209 09:48:41.874821 6920 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx\\\\nI1209 09:48:41.874844 6920 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1209 09:48:41.874869 6920 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}\\\\nI1209 09:48:41.874881 6920 services_controller.go:360] Finished syncing service olm-operator-metrics on namespace openshift-operator-lifecycle-manager for network=default : 7.206478ms\\\\nI1209 09:48:41.874892 6920 services_controller.go:356] Processing sync for service openshift-dns/dns-default for network=default\\\\nF1209 09:48:41.874931 6920 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:48:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vvxck_openshift-ovn-kubernetes(13620d58-6925-48c7-bbc7-f34e92922480)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.654433 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daad2ae578b4f337ceb3298b012b0262903cebe4e9e9584513cf2f31fbcec3c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37860b2742b9486e06919b4bb4e7e4631ff5905eea837ef17c1d2f8d573da18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gcrqx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.666293 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.666344 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.666354 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.666372 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.666381 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:43Z","lastTransitionTime":"2025-12-09T09:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.769051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.769084 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.769093 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.769106 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.769116 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:43Z","lastTransitionTime":"2025-12-09T09:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.871434 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.871484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.871496 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.871512 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.871524 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:43Z","lastTransitionTime":"2025-12-09T09:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.910224 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:43 crc kubenswrapper[4824]: E1209 09:48:43.910378 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.924165 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0e3820ffa2ddaaf2e6ccf12288f63187d08eaf2ea0bc868e19a3018a0d92357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.936201 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7efb0cf4795c447fbf546c9f99ea8424442dd16a0fd684cf900d5865d53544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da54c40b4abbad5fe19ee4d98b4e38446da7a58d0c9594ee2a31ffb5c0c87928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.946893 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sxd45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40095fcc-84f8-4c61-8255-f76878bdc3bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43f14a20202430d38963b4b9987fe18bfa26fd217231933ccdc6cb99b0b3de49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4s87f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sxd45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.960919 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2dd916-62ee-4dca-964d-fb859475a9d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1e6cfeb997cafb7ec7f6d829e1ff2e2d95e4b343c422194a5290d222ea53a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2add2743e390cf2b91c9b020de38fc6df392fee9e5aa1d59bd1ca9471d9d8d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17601a452dc4369a533806384d58f733e396e255f5741b38004cae98a44cfe55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eb98ce1943d5692f7e6d51befb0c05115ee403be72134f40f651a80bf0a378b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8eb98ce1943d5692f7e6d51befb0c05115ee403be72134f40f651a80bf0a378b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.974417 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.974463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.974473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.974498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.974509 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:43Z","lastTransitionTime":"2025-12-09T09:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.975951 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:43 crc kubenswrapper[4824]: I1209 09:48:43.990184 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b57098a174d766f3b6d289f0d765188da1e6b503cf25e10a66dd4b6018b20dde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:43Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.004964 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pvksg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74aa3cd7-0abc-4fc9-9697-e23b693a7611\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0ee66e89363676597efd13d02245ce8c9a9af288c04f2ed9b7ae8bed31da259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:48:31Z\\\",\\\"message\\\":\\\"2025-12-09T09:47:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c3558da8-38fa-42ec-b5cd-4f174f21b210\\\\n2025-12-09T09:47:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c3558da8-38fa-42ec-b5cd-4f174f21b210 to /host/opt/cni/bin/\\\\n2025-12-09T09:47:46Z [verbose] multus-daemon started\\\\n2025-12-09T09:47:46Z [verbose] Readiness Indicator file check\\\\n2025-12-09T09:48:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs2pc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pvksg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.016604 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.030694 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e41bcf8-bacf-4eb1-a391-bbd86adedb32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://161db63a43e5b9c4b2f74e575168916be5ebb688eaa59012003d88ed4669b181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660496dd9f31918ac2dd4facaf91054acd006e2d45781da1db895cc5b105853b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579c0893f6574908a66c4c3c1f60a750cdbfb22a5a148bf313bbf738cb50676f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4bc3a62ae8edaf4c904253e5d0988c75a7c2977224b1b24c9298286cf6247f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45acac5938a46337259f853a3091a18f05606eb8ee471d7da919dc225299053d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62ae32280b4dfeb9fcf5556c0ee189b6ab7c245241b65eff7e60ac572e1d03fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da399da8d09d72dfd5c7f838856cf7eeb7b23824c4025df54da4395143e8866\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcwxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-w8sjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.050256 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13620d58-6925-48c7-bbc7-f34e92922480\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55659b94cb673db2b20bd35a15232443b44882962e7731116b74c418ea5d251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55659b94cb673db2b20bd35a15232443b44882962e7731116b74c418ea5d251e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T09:48:42Z\\\",\\\"message\\\":\\\"ceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1209 09:48:41.874805 6920 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx in node crc\\\\nI1209 09:48:41.874813 6920 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx after 0 failed attempt(s)\\\\nI1209 09:48:41.874821 6920 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx\\\\nI1209 09:48:41.874844 6920 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1209 09:48:41.874869 6920 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}\\\\nI1209 09:48:41.874881 6920 services_controller.go:360] Finished syncing service olm-operator-metrics on namespace openshift-operator-lifecycle-manager for network=default : 7.206478ms\\\\nI1209 09:48:41.874892 6920 services_controller.go:356] Processing sync for service openshift-dns/dns-default for network=default\\\\nF1209 09:48:41.874931 6920 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:48:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vvxck_openshift-ovn-kubernetes(13620d58-6925-48c7-bbc7-f34e92922480)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrd4m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vvxck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.064275 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1fdd4fd7-3d1c-475e-a225-6b0a9145ffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daad2ae578b4f337ceb3298b012b0262903cebe4e9e9584513cf2f31fbcec3c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37860b2742b9486e06919b4bb4e7e4631ff5905eea837ef17c1d2f8d573da18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-256nl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gcrqx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.077852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.077916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.077931 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.077957 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.077977 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:44Z","lastTransitionTime":"2025-12-09T09:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.078328 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5tn52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psr7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5tn52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.093707 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zhqgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4d3c84-c8c8-4c5b-8e15-95ab59ea77cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4721617c88775be8645f44d62924ef15e13b4e72fd336eb90b9a08f139f1deb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lk9bt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zhqgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.112295 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee3c86bf-f246-42cb-adb6-900e3d735166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27fae0d9fdc6961f9641567b5684aa7266f84af2aa03ce7cf7a4ba079272e07f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5wtmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dth8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.136936 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a675c175-8e33-422c-b361-6cd873bd82d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81e1e5effeceb04c4f0d31bd6d8db4e42611d2a168bc455424f1660c002547a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ee2908650e4147fc4da10c5719cbd0da2f5e0c7b6afd69a8084032261133650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a66a31086a2115100d1753b05d6c27eb9b232a53f4b478a50bf02c31e14252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1d058aa00a30319fb6a21e95b79c21ab4623b5306eb42e3f22e2db7bcae59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b08e63909e92a503d68d63cd08e7c69e36cd0ae180b2b3dfdf9cdc3d5aeb5f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21573b549f4704275a9f6f387063f914b32da6adad85a086ba05e1c0ea467e3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f40a5d5d8104d4cfb2fba0f723e3a25cbdc0de8f6d1adefcf6ea0435378d52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://058fb239cde6b4a80a35f769ed6151f43bfe92df93128184d463c2e814a803f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.154587 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9ca83da-bb34-4480-8fa1-9cca30abfb02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 09:47:37.564025 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 09:47:37.565315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3377060745/tls.crt::/tmp/serving-cert-3377060745/tls.key\\\\\\\"\\\\nI1209 09:47:43.019543 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 09:47:43.022660 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 09:47:43.023102 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 09:47:43.023209 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 09:47:43.023255 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 09:47:43.029809 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 09:47:43.030559 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030568 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 09:47:43.030573 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 09:47:43.030577 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 09:47:43.030581 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1209 09:47:43.029865 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 09:47:43.030619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 09:47:43.033599 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T09:47:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.176583 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8679f1aa-2458-4c97-b132-c289c8584519\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932dc72c0ad07cae1d98b18e2cdaf60252e3d14e03e5eff5f533bddf7540de68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5ee7e0988679424d8c0ae8e4338794af407bba8365b651389cdc33625b26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96f7a44969ad38de369a0876998dabd230e3a053c53f00bfc05bd087b8953154\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T09:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T09:47:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.181938 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.181989 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.182001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.182023 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.182035 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:44Z","lastTransitionTime":"2025-12-09T09:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.197168 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T09:47:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T09:48:44Z is after 2025-08-24T17:21:41Z" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.286118 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.286448 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.286554 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.286620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.286695 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:44Z","lastTransitionTime":"2025-12-09T09:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.389457 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.389495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.389505 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.389523 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.389534 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:44Z","lastTransitionTime":"2025-12-09T09:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.492594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.492648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.492660 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.492681 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.492694 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:44Z","lastTransitionTime":"2025-12-09T09:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.595258 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.595305 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.595320 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.595342 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.595353 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:44Z","lastTransitionTime":"2025-12-09T09:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.697875 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.697914 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.697923 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.697941 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.697951 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:44Z","lastTransitionTime":"2025-12-09T09:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.799983 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.800314 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.800410 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.800501 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.800576 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:44Z","lastTransitionTime":"2025-12-09T09:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.902890 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.902943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.902954 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.902970 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.902983 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:44Z","lastTransitionTime":"2025-12-09T09:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.910386 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.910417 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:44 crc kubenswrapper[4824]: E1209 09:48:44.910546 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:44 crc kubenswrapper[4824]: I1209 09:48:44.910424 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:44 crc kubenswrapper[4824]: E1209 09:48:44.910697 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:44 crc kubenswrapper[4824]: E1209 09:48:44.910852 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.006003 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.006077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.006090 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.006110 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.006122 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:45Z","lastTransitionTime":"2025-12-09T09:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.108840 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.108889 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.108902 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.108920 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.108931 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:45Z","lastTransitionTime":"2025-12-09T09:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.211929 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.212008 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.212032 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.212062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.212086 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:45Z","lastTransitionTime":"2025-12-09T09:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.315399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.315688 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.315852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.315983 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.316066 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:45Z","lastTransitionTime":"2025-12-09T09:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.418215 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.418256 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.418264 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.418282 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.418294 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:45Z","lastTransitionTime":"2025-12-09T09:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.520636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.520670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.520679 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.520693 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.520703 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:45Z","lastTransitionTime":"2025-12-09T09:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.622973 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.623013 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.623023 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.623042 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.623056 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:45Z","lastTransitionTime":"2025-12-09T09:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.725638 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.725683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.725691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.725708 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.725720 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:45Z","lastTransitionTime":"2025-12-09T09:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.828696 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.828769 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.828801 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.828828 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.828842 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:45Z","lastTransitionTime":"2025-12-09T09:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.909474 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:45 crc kubenswrapper[4824]: E1209 09:48:45.909775 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.931889 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.931931 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.931942 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.931958 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:45 crc kubenswrapper[4824]: I1209 09:48:45.931971 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:45Z","lastTransitionTime":"2025-12-09T09:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.034482 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.034556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.034572 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.034591 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.034603 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:46Z","lastTransitionTime":"2025-12-09T09:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.137917 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.137989 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.138005 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.138149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.138168 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:46Z","lastTransitionTime":"2025-12-09T09:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.241351 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.241432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.241447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.241470 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.241486 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:46Z","lastTransitionTime":"2025-12-09T09:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.345462 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.345554 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.345566 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.345583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.345594 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:46Z","lastTransitionTime":"2025-12-09T09:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.449264 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.449317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.449327 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.449347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.449359 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:46Z","lastTransitionTime":"2025-12-09T09:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.553590 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.553639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.553652 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.553673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.553686 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:46Z","lastTransitionTime":"2025-12-09T09:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.657402 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.657461 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.657473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.657490 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.657501 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:46Z","lastTransitionTime":"2025-12-09T09:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.761366 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.761452 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.761465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.761486 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.761499 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:46Z","lastTransitionTime":"2025-12-09T09:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.768950 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:48:46 crc kubenswrapper[4824]: E1209 09:48:46.769101 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:50.769071297 +0000 UTC m=+147.103576024 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.863818 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.863872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.863882 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.863903 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.863916 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:46Z","lastTransitionTime":"2025-12-09T09:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.870368 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.870432 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.870461 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.870490 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:46 crc kubenswrapper[4824]: E1209 09:48:46.870576 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 09:48:46 crc kubenswrapper[4824]: E1209 09:48:46.870609 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 09:48:46 crc kubenswrapper[4824]: E1209 09:48:46.870612 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 09:48:46 crc kubenswrapper[4824]: E1209 09:48:46.870668 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 09:48:46 crc kubenswrapper[4824]: E1209 09:48:46.870685 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:48:46 crc kubenswrapper[4824]: E1209 09:48:46.870670 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 09:49:50.870642524 +0000 UTC m=+147.205147191 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 09:48:46 crc kubenswrapper[4824]: E1209 09:48:46.870830 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 09:49:50.870766858 +0000 UTC m=+147.205271605 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 09:48:46 crc kubenswrapper[4824]: E1209 09:48:46.870837 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 09:48:46 crc kubenswrapper[4824]: E1209 09:48:46.870901 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 09:48:46 crc kubenswrapper[4824]: E1209 09:48:46.870918 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:48:46 crc kubenswrapper[4824]: E1209 09:48:46.870858 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 09:49:50.870851311 +0000 UTC m=+147.205356098 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:48:46 crc kubenswrapper[4824]: E1209 09:48:46.871030 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 09:49:50.871001664 +0000 UTC m=+147.205506421 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.910511 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.910608 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.910528 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:46 crc kubenswrapper[4824]: E1209 09:48:46.910719 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:46 crc kubenswrapper[4824]: E1209 09:48:46.910989 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:46 crc kubenswrapper[4824]: E1209 09:48:46.911139 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.966832 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.966878 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.966888 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.966903 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:46 crc kubenswrapper[4824]: I1209 09:48:46.966913 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:46Z","lastTransitionTime":"2025-12-09T09:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.071026 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.071092 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.071166 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.071200 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.071223 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:47Z","lastTransitionTime":"2025-12-09T09:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.174385 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.174419 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.174427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.174441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.174468 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:47Z","lastTransitionTime":"2025-12-09T09:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.276164 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.276212 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.276227 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.276244 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.276255 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:47Z","lastTransitionTime":"2025-12-09T09:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.379226 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.379276 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.379287 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.379304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.379316 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:47Z","lastTransitionTime":"2025-12-09T09:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.482424 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.482464 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.482474 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.482492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.482503 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:47Z","lastTransitionTime":"2025-12-09T09:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.585119 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.586106 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.586121 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.586140 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.586155 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:47Z","lastTransitionTime":"2025-12-09T09:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.689520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.689610 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.689637 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.689670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.689693 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:47Z","lastTransitionTime":"2025-12-09T09:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.792190 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.792241 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.792253 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.792270 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.792283 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:47Z","lastTransitionTime":"2025-12-09T09:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.894462 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.894516 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.894530 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.894549 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.894564 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:47Z","lastTransitionTime":"2025-12-09T09:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.909917 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:47 crc kubenswrapper[4824]: E1209 09:48:47.910167 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.996940 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.997001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.997015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.997048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:47 crc kubenswrapper[4824]: I1209 09:48:47.997060 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:47Z","lastTransitionTime":"2025-12-09T09:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.098933 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.098972 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.098981 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.098995 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.099005 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:48Z","lastTransitionTime":"2025-12-09T09:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.200626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.200666 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.200684 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.200704 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.200718 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:48Z","lastTransitionTime":"2025-12-09T09:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.309530 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.309660 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.309678 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.309704 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.309722 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:48Z","lastTransitionTime":"2025-12-09T09:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.412163 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.412492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.412502 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.412515 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.412525 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:48Z","lastTransitionTime":"2025-12-09T09:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.483747 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.483860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.483889 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.483917 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.483937 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T09:48:48Z","lastTransitionTime":"2025-12-09T09:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.526469 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-p4wrc"] Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.526846 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p4wrc" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.533230 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.533726 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.533738 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.533750 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.561771 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-zhqgl" podStartSLOduration=64.561746738 podStartE2EDuration="1m4.561746738s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:48:48.56147972 +0000 UTC m=+84.895984407" watchObservedRunningTime="2025-12-09 09:48:48.561746738 +0000 UTC m=+84.896251405" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.586415 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4a344b84-1214-426c-ba63-2a6cb971d5c8-service-ca\") pod \"cluster-version-operator-5c965bbfc6-p4wrc\" (UID: \"4a344b84-1214-426c-ba63-2a6cb971d5c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p4wrc" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.586474 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/4a344b84-1214-426c-ba63-2a6cb971d5c8-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-p4wrc\" (UID: \"4a344b84-1214-426c-ba63-2a6cb971d5c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p4wrc" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.586568 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/4a344b84-1214-426c-ba63-2a6cb971d5c8-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-p4wrc\" (UID: \"4a344b84-1214-426c-ba63-2a6cb971d5c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p4wrc" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.586732 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4a344b84-1214-426c-ba63-2a6cb971d5c8-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-p4wrc\" (UID: \"4a344b84-1214-426c-ba63-2a6cb971d5c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p4wrc" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.586775 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a344b84-1214-426c-ba63-2a6cb971d5c8-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-p4wrc\" (UID: \"4a344b84-1214-426c-ba63-2a6cb971d5c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p4wrc" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.596160 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podStartSLOduration=64.596141637 podStartE2EDuration="1m4.596141637s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:48:48.57223443 +0000 UTC m=+84.906739127" watchObservedRunningTime="2025-12-09 09:48:48.596141637 +0000 UTC m=+84.930646304" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.596433 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=63.596428106 podStartE2EDuration="1m3.596428106s" podCreationTimestamp="2025-12-09 09:47:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:48:48.595651881 +0000 UTC m=+84.930156548" watchObservedRunningTime="2025-12-09 09:48:48.596428106 +0000 UTC m=+84.930932773" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.637591 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=64.637565528 podStartE2EDuration="1m4.637565528s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:48:48.620929922 +0000 UTC m=+84.955434589" watchObservedRunningTime="2025-12-09 09:48:48.637565528 +0000 UTC m=+84.972070205" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.638003 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=65.637997902 podStartE2EDuration="1m5.637997902s" podCreationTimestamp="2025-12-09 09:47:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:48:48.637459786 +0000 UTC m=+84.971964473" watchObservedRunningTime="2025-12-09 09:48:48.637997902 +0000 UTC m=+84.972502569" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.687829 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4a344b84-1214-426c-ba63-2a6cb971d5c8-service-ca\") pod \"cluster-version-operator-5c965bbfc6-p4wrc\" (UID: \"4a344b84-1214-426c-ba63-2a6cb971d5c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p4wrc" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.687872 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/4a344b84-1214-426c-ba63-2a6cb971d5c8-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-p4wrc\" (UID: \"4a344b84-1214-426c-ba63-2a6cb971d5c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p4wrc" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.687913 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/4a344b84-1214-426c-ba63-2a6cb971d5c8-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-p4wrc\" (UID: \"4a344b84-1214-426c-ba63-2a6cb971d5c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p4wrc" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.687947 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a344b84-1214-426c-ba63-2a6cb971d5c8-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-p4wrc\" (UID: \"4a344b84-1214-426c-ba63-2a6cb971d5c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p4wrc" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.687968 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4a344b84-1214-426c-ba63-2a6cb971d5c8-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-p4wrc\" (UID: \"4a344b84-1214-426c-ba63-2a6cb971d5c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p4wrc" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.688167 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/4a344b84-1214-426c-ba63-2a6cb971d5c8-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-p4wrc\" (UID: \"4a344b84-1214-426c-ba63-2a6cb971d5c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p4wrc" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.688300 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/4a344b84-1214-426c-ba63-2a6cb971d5c8-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-p4wrc\" (UID: \"4a344b84-1214-426c-ba63-2a6cb971d5c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p4wrc" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.688422 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-sxd45" podStartSLOduration=64.688397008 podStartE2EDuration="1m4.688397008s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:48:48.678358401 +0000 UTC m=+85.012863068" watchObservedRunningTime="2025-12-09 09:48:48.688397008 +0000 UTC m=+85.022901675" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.689002 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4a344b84-1214-426c-ba63-2a6cb971d5c8-service-ca\") pod \"cluster-version-operator-5c965bbfc6-p4wrc\" (UID: \"4a344b84-1214-426c-ba63-2a6cb971d5c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p4wrc" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.694919 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a344b84-1214-426c-ba63-2a6cb971d5c8-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-p4wrc\" (UID: \"4a344b84-1214-426c-ba63-2a6cb971d5c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p4wrc" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.701948 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=32.701930157 podStartE2EDuration="32.701930157s" podCreationTimestamp="2025-12-09 09:48:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:48:48.701477003 +0000 UTC m=+85.035981700" watchObservedRunningTime="2025-12-09 09:48:48.701930157 +0000 UTC m=+85.036434824" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.714320 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4a344b84-1214-426c-ba63-2a6cb971d5c8-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-p4wrc\" (UID: \"4a344b84-1214-426c-ba63-2a6cb971d5c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p4wrc" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.749618 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-pvksg" podStartSLOduration=64.749599026 podStartE2EDuration="1m4.749599026s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:48:48.749421242 +0000 UTC m=+85.083925929" watchObservedRunningTime="2025-12-09 09:48:48.749599026 +0000 UTC m=+85.084103693" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.784731 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-w8sjn" podStartSLOduration=64.784710689 podStartE2EDuration="1m4.784710689s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:48:48.783870312 +0000 UTC m=+85.118374989" watchObservedRunningTime="2025-12-09 09:48:48.784710689 +0000 UTC m=+85.119215346" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.818624 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gcrqx" podStartSLOduration=64.818603072 podStartE2EDuration="1m4.818603072s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:48:48.817928561 +0000 UTC m=+85.152433248" watchObservedRunningTime="2025-12-09 09:48:48.818603072 +0000 UTC m=+85.153107739" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.841153 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p4wrc" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.909606 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.909607 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:48 crc kubenswrapper[4824]: I1209 09:48:48.909754 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:48 crc kubenswrapper[4824]: E1209 09:48:48.910572 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:48 crc kubenswrapper[4824]: E1209 09:48:48.910972 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:48 crc kubenswrapper[4824]: E1209 09:48:48.911096 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:49 crc kubenswrapper[4824]: I1209 09:48:49.412859 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p4wrc" event={"ID":"4a344b84-1214-426c-ba63-2a6cb971d5c8","Type":"ContainerStarted","Data":"155339618aab5906508461ed42a4fcb6be80645d02651db3a213a407a9d950af"} Dec 09 09:48:49 crc kubenswrapper[4824]: I1209 09:48:49.413337 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p4wrc" event={"ID":"4a344b84-1214-426c-ba63-2a6cb971d5c8","Type":"ContainerStarted","Data":"9907102c0c9f4ae9a3d89f18c35963572b817b337190b395ed293aa215891c14"} Dec 09 09:48:49 crc kubenswrapper[4824]: I1209 09:48:49.429758 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p4wrc" podStartSLOduration=65.429729036 podStartE2EDuration="1m5.429729036s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:48:49.429303603 +0000 UTC m=+85.763808280" watchObservedRunningTime="2025-12-09 09:48:49.429729036 +0000 UTC m=+85.764233713" Dec 09 09:48:49 crc kubenswrapper[4824]: I1209 09:48:49.910418 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:49 crc kubenswrapper[4824]: E1209 09:48:49.910617 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:50 crc kubenswrapper[4824]: I1209 09:48:50.910217 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:50 crc kubenswrapper[4824]: I1209 09:48:50.910305 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:50 crc kubenswrapper[4824]: I1209 09:48:50.910223 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:50 crc kubenswrapper[4824]: E1209 09:48:50.910444 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:50 crc kubenswrapper[4824]: E1209 09:48:50.910753 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:50 crc kubenswrapper[4824]: E1209 09:48:50.910982 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:51 crc kubenswrapper[4824]: I1209 09:48:51.910310 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:51 crc kubenswrapper[4824]: E1209 09:48:51.910451 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:52 crc kubenswrapper[4824]: I1209 09:48:52.909986 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:52 crc kubenswrapper[4824]: I1209 09:48:52.910049 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:52 crc kubenswrapper[4824]: I1209 09:48:52.910059 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:52 crc kubenswrapper[4824]: E1209 09:48:52.910162 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:52 crc kubenswrapper[4824]: E1209 09:48:52.910249 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:52 crc kubenswrapper[4824]: E1209 09:48:52.910309 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:53 crc kubenswrapper[4824]: I1209 09:48:53.909583 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:53 crc kubenswrapper[4824]: E1209 09:48:53.912929 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:54 crc kubenswrapper[4824]: I1209 09:48:54.910501 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:54 crc kubenswrapper[4824]: I1209 09:48:54.910591 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:54 crc kubenswrapper[4824]: E1209 09:48:54.910761 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:54 crc kubenswrapper[4824]: I1209 09:48:54.911026 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:54 crc kubenswrapper[4824]: E1209 09:48:54.911221 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:54 crc kubenswrapper[4824]: E1209 09:48:54.911365 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:55 crc kubenswrapper[4824]: I1209 09:48:55.909944 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:55 crc kubenswrapper[4824]: E1209 09:48:55.910168 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:56 crc kubenswrapper[4824]: I1209 09:48:56.910047 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:56 crc kubenswrapper[4824]: I1209 09:48:56.910103 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:56 crc kubenswrapper[4824]: E1209 09:48:56.910711 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:56 crc kubenswrapper[4824]: I1209 09:48:56.910199 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:56 crc kubenswrapper[4824]: E1209 09:48:56.910526 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:56 crc kubenswrapper[4824]: E1209 09:48:56.911182 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:56 crc kubenswrapper[4824]: I1209 09:48:56.925128 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 09 09:48:57 crc kubenswrapper[4824]: I1209 09:48:57.910476 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:57 crc kubenswrapper[4824]: I1209 09:48:57.911745 4824 scope.go:117] "RemoveContainer" containerID="55659b94cb673db2b20bd35a15232443b44882962e7731116b74c418ea5d251e" Dec 09 09:48:57 crc kubenswrapper[4824]: E1209 09:48:57.912048 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vvxck_openshift-ovn-kubernetes(13620d58-6925-48c7-bbc7-f34e92922480)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" podUID="13620d58-6925-48c7-bbc7-f34e92922480" Dec 09 09:48:57 crc kubenswrapper[4824]: E1209 09:48:57.912040 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:48:58 crc kubenswrapper[4824]: I1209 09:48:58.909875 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:48:58 crc kubenswrapper[4824]: I1209 09:48:58.909931 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:48:58 crc kubenswrapper[4824]: E1209 09:48:58.910465 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:48:58 crc kubenswrapper[4824]: E1209 09:48:58.910319 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:48:58 crc kubenswrapper[4824]: I1209 09:48:58.909985 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:48:58 crc kubenswrapper[4824]: E1209 09:48:58.910744 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:48:59 crc kubenswrapper[4824]: I1209 09:48:59.909683 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:48:59 crc kubenswrapper[4824]: E1209 09:48:59.910323 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:49:00 crc kubenswrapper[4824]: I1209 09:49:00.910196 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:49:00 crc kubenswrapper[4824]: I1209 09:49:00.910355 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:49:00 crc kubenswrapper[4824]: E1209 09:49:00.910466 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:49:00 crc kubenswrapper[4824]: I1209 09:49:00.910869 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:49:00 crc kubenswrapper[4824]: E1209 09:49:00.910916 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:49:00 crc kubenswrapper[4824]: E1209 09:49:00.911017 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:49:01 crc kubenswrapper[4824]: I1209 09:49:01.910355 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:49:01 crc kubenswrapper[4824]: E1209 09:49:01.910518 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:49:02 crc kubenswrapper[4824]: I1209 09:49:02.535921 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs\") pod \"network-metrics-daemon-5tn52\" (UID: \"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\") " pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:49:02 crc kubenswrapper[4824]: E1209 09:49:02.536195 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 09:49:02 crc kubenswrapper[4824]: E1209 09:49:02.536325 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs podName:fcbce9ea-3d55-4e09-b5bc-d5bb8c472219 nodeName:}" failed. No retries permitted until 2025-12-09 09:50:06.536285472 +0000 UTC m=+162.870790169 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs") pod "network-metrics-daemon-5tn52" (UID: "fcbce9ea-3d55-4e09-b5bc-d5bb8c472219") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 09:49:02 crc kubenswrapper[4824]: I1209 09:49:02.909624 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:49:02 crc kubenswrapper[4824]: I1209 09:49:02.909690 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:49:02 crc kubenswrapper[4824]: I1209 09:49:02.909690 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:49:02 crc kubenswrapper[4824]: E1209 09:49:02.909828 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:49:02 crc kubenswrapper[4824]: E1209 09:49:02.910214 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:49:02 crc kubenswrapper[4824]: E1209 09:49:02.910455 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:49:03 crc kubenswrapper[4824]: I1209 09:49:03.909517 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:49:03 crc kubenswrapper[4824]: E1209 09:49:03.911599 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:49:03 crc kubenswrapper[4824]: I1209 09:49:03.924587 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=7.9245643359999995 podStartE2EDuration="7.924564336s" podCreationTimestamp="2025-12-09 09:48:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:03.92440675 +0000 UTC m=+100.258911427" watchObservedRunningTime="2025-12-09 09:49:03.924564336 +0000 UTC m=+100.259069003" Dec 09 09:49:04 crc kubenswrapper[4824]: I1209 09:49:04.909396 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:49:04 crc kubenswrapper[4824]: I1209 09:49:04.909404 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:49:04 crc kubenswrapper[4824]: E1209 09:49:04.909570 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:49:04 crc kubenswrapper[4824]: E1209 09:49:04.909638 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:49:04 crc kubenswrapper[4824]: I1209 09:49:04.909435 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:49:04 crc kubenswrapper[4824]: E1209 09:49:04.909741 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:49:05 crc kubenswrapper[4824]: I1209 09:49:05.912361 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:49:05 crc kubenswrapper[4824]: E1209 09:49:05.912487 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:49:06 crc kubenswrapper[4824]: I1209 09:49:06.909941 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:49:06 crc kubenswrapper[4824]: I1209 09:49:06.910035 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:49:06 crc kubenswrapper[4824]: E1209 09:49:06.910079 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:49:06 crc kubenswrapper[4824]: I1209 09:49:06.910142 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:49:06 crc kubenswrapper[4824]: E1209 09:49:06.910195 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:49:06 crc kubenswrapper[4824]: E1209 09:49:06.910333 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:49:07 crc kubenswrapper[4824]: I1209 09:49:07.910368 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:49:07 crc kubenswrapper[4824]: E1209 09:49:07.910668 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:49:08 crc kubenswrapper[4824]: I1209 09:49:08.910154 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:49:08 crc kubenswrapper[4824]: I1209 09:49:08.910130 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:49:08 crc kubenswrapper[4824]: I1209 09:49:08.910170 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:49:08 crc kubenswrapper[4824]: E1209 09:49:08.910306 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:49:08 crc kubenswrapper[4824]: E1209 09:49:08.910420 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:49:08 crc kubenswrapper[4824]: E1209 09:49:08.910527 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:49:09 crc kubenswrapper[4824]: I1209 09:49:09.910219 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:49:09 crc kubenswrapper[4824]: E1209 09:49:09.910699 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:49:10 crc kubenswrapper[4824]: I1209 09:49:10.909500 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:49:10 crc kubenswrapper[4824]: I1209 09:49:10.909552 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:49:10 crc kubenswrapper[4824]: E1209 09:49:10.909615 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:49:10 crc kubenswrapper[4824]: I1209 09:49:10.909552 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:49:10 crc kubenswrapper[4824]: E1209 09:49:10.909690 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:49:10 crc kubenswrapper[4824]: E1209 09:49:10.909777 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:49:11 crc kubenswrapper[4824]: I1209 09:49:11.911986 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:49:11 crc kubenswrapper[4824]: E1209 09:49:11.912206 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:49:12 crc kubenswrapper[4824]: I1209 09:49:12.909970 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:49:12 crc kubenswrapper[4824]: E1209 09:49:12.910394 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:49:12 crc kubenswrapper[4824]: I1209 09:49:12.910139 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:49:12 crc kubenswrapper[4824]: E1209 09:49:12.910650 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:49:12 crc kubenswrapper[4824]: I1209 09:49:12.910717 4824 scope.go:117] "RemoveContainer" containerID="55659b94cb673db2b20bd35a15232443b44882962e7731116b74c418ea5d251e" Dec 09 09:49:12 crc kubenswrapper[4824]: I1209 09:49:12.910126 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:49:12 crc kubenswrapper[4824]: E1209 09:49:12.910882 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vvxck_openshift-ovn-kubernetes(13620d58-6925-48c7-bbc7-f34e92922480)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" podUID="13620d58-6925-48c7-bbc7-f34e92922480" Dec 09 09:49:12 crc kubenswrapper[4824]: E1209 09:49:12.910990 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:49:13 crc kubenswrapper[4824]: I1209 09:49:13.909908 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:49:13 crc kubenswrapper[4824]: E1209 09:49:13.911005 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:49:14 crc kubenswrapper[4824]: I1209 09:49:14.909828 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:49:14 crc kubenswrapper[4824]: I1209 09:49:14.909828 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:49:14 crc kubenswrapper[4824]: I1209 09:49:14.910847 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:49:14 crc kubenswrapper[4824]: E1209 09:49:14.911132 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:49:14 crc kubenswrapper[4824]: E1209 09:49:14.911408 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:49:14 crc kubenswrapper[4824]: E1209 09:49:14.911499 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:49:15 crc kubenswrapper[4824]: I1209 09:49:15.910463 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:49:15 crc kubenswrapper[4824]: E1209 09:49:15.910681 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:49:16 crc kubenswrapper[4824]: I1209 09:49:16.910111 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:49:16 crc kubenswrapper[4824]: I1209 09:49:16.910124 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:49:16 crc kubenswrapper[4824]: I1209 09:49:16.910298 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:49:16 crc kubenswrapper[4824]: E1209 09:49:16.910509 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:49:16 crc kubenswrapper[4824]: E1209 09:49:16.910630 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:49:16 crc kubenswrapper[4824]: E1209 09:49:16.910727 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:49:17 crc kubenswrapper[4824]: I1209 09:49:17.910631 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:49:17 crc kubenswrapper[4824]: E1209 09:49:17.910886 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:49:18 crc kubenswrapper[4824]: I1209 09:49:18.506164 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-pvksg_74aa3cd7-0abc-4fc9-9697-e23b693a7611/kube-multus/1.log" Dec 09 09:49:18 crc kubenswrapper[4824]: I1209 09:49:18.506895 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-pvksg_74aa3cd7-0abc-4fc9-9697-e23b693a7611/kube-multus/0.log" Dec 09 09:49:18 crc kubenswrapper[4824]: I1209 09:49:18.506960 4824 generic.go:334] "Generic (PLEG): container finished" podID="74aa3cd7-0abc-4fc9-9697-e23b693a7611" containerID="c0ee66e89363676597efd13d02245ce8c9a9af288c04f2ed9b7ae8bed31da259" exitCode=1 Dec 09 09:49:18 crc kubenswrapper[4824]: I1209 09:49:18.507014 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pvksg" event={"ID":"74aa3cd7-0abc-4fc9-9697-e23b693a7611","Type":"ContainerDied","Data":"c0ee66e89363676597efd13d02245ce8c9a9af288c04f2ed9b7ae8bed31da259"} Dec 09 09:49:18 crc kubenswrapper[4824]: I1209 09:49:18.507105 4824 scope.go:117] "RemoveContainer" containerID="cccf72fbf8aad26ef5ad501211394fe9505940dc966f17923ec4f28038db36de" Dec 09 09:49:18 crc kubenswrapper[4824]: I1209 09:49:18.507426 4824 scope.go:117] "RemoveContainer" containerID="c0ee66e89363676597efd13d02245ce8c9a9af288c04f2ed9b7ae8bed31da259" Dec 09 09:49:18 crc kubenswrapper[4824]: E1209 09:49:18.507661 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-pvksg_openshift-multus(74aa3cd7-0abc-4fc9-9697-e23b693a7611)\"" pod="openshift-multus/multus-pvksg" podUID="74aa3cd7-0abc-4fc9-9697-e23b693a7611" Dec 09 09:49:18 crc kubenswrapper[4824]: I1209 09:49:18.910406 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:49:18 crc kubenswrapper[4824]: I1209 09:49:18.910474 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:49:18 crc kubenswrapper[4824]: I1209 09:49:18.910430 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:49:18 crc kubenswrapper[4824]: E1209 09:49:18.910663 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:49:18 crc kubenswrapper[4824]: E1209 09:49:18.910876 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:49:18 crc kubenswrapper[4824]: E1209 09:49:18.911016 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:49:19 crc kubenswrapper[4824]: I1209 09:49:19.512653 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-pvksg_74aa3cd7-0abc-4fc9-9697-e23b693a7611/kube-multus/1.log" Dec 09 09:49:19 crc kubenswrapper[4824]: I1209 09:49:19.909590 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:49:19 crc kubenswrapper[4824]: E1209 09:49:19.909853 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:49:20 crc kubenswrapper[4824]: I1209 09:49:20.909739 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:49:20 crc kubenswrapper[4824]: E1209 09:49:20.910838 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:49:20 crc kubenswrapper[4824]: I1209 09:49:20.909762 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:49:20 crc kubenswrapper[4824]: E1209 09:49:20.911076 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:49:20 crc kubenswrapper[4824]: I1209 09:49:20.909762 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:49:20 crc kubenswrapper[4824]: E1209 09:49:20.911298 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:49:21 crc kubenswrapper[4824]: I1209 09:49:21.910313 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:49:21 crc kubenswrapper[4824]: E1209 09:49:21.910505 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:49:22 crc kubenswrapper[4824]: I1209 09:49:22.909726 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:49:22 crc kubenswrapper[4824]: I1209 09:49:22.909800 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:49:22 crc kubenswrapper[4824]: I1209 09:49:22.909745 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:49:22 crc kubenswrapper[4824]: E1209 09:49:22.909888 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:49:22 crc kubenswrapper[4824]: E1209 09:49:22.909941 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:49:22 crc kubenswrapper[4824]: E1209 09:49:22.910063 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:49:23 crc kubenswrapper[4824]: E1209 09:49:23.870031 4824 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 09 09:49:23 crc kubenswrapper[4824]: I1209 09:49:23.910317 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:49:23 crc kubenswrapper[4824]: E1209 09:49:23.911860 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:49:24 crc kubenswrapper[4824]: E1209 09:49:24.100076 4824 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 09:49:24 crc kubenswrapper[4824]: I1209 09:49:24.909595 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:49:24 crc kubenswrapper[4824]: E1209 09:49:24.909745 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:49:24 crc kubenswrapper[4824]: I1209 09:49:24.909834 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:49:24 crc kubenswrapper[4824]: E1209 09:49:24.910032 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:49:24 crc kubenswrapper[4824]: I1209 09:49:24.910310 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:49:24 crc kubenswrapper[4824]: E1209 09:49:24.910439 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:49:25 crc kubenswrapper[4824]: I1209 09:49:25.912240 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:49:25 crc kubenswrapper[4824]: E1209 09:49:25.912384 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:49:26 crc kubenswrapper[4824]: I1209 09:49:26.910458 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:49:26 crc kubenswrapper[4824]: I1209 09:49:26.910511 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:49:26 crc kubenswrapper[4824]: E1209 09:49:26.910602 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:49:26 crc kubenswrapper[4824]: I1209 09:49:26.910700 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:49:26 crc kubenswrapper[4824]: E1209 09:49:26.910917 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:49:26 crc kubenswrapper[4824]: E1209 09:49:26.910945 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:49:26 crc kubenswrapper[4824]: I1209 09:49:26.912008 4824 scope.go:117] "RemoveContainer" containerID="55659b94cb673db2b20bd35a15232443b44882962e7731116b74c418ea5d251e" Dec 09 09:49:27 crc kubenswrapper[4824]: I1209 09:49:27.542383 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vvxck_13620d58-6925-48c7-bbc7-f34e92922480/ovnkube-controller/3.log" Dec 09 09:49:27 crc kubenswrapper[4824]: I1209 09:49:27.546202 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerStarted","Data":"78f6ecf0c19cc7fb61096fcb23d74d6682beca46786235c4a23199197f6757a5"} Dec 09 09:49:27 crc kubenswrapper[4824]: I1209 09:49:27.546688 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:49:27 crc kubenswrapper[4824]: I1209 09:49:27.585347 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" podStartSLOduration=103.585329298 podStartE2EDuration="1m43.585329298s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:27.583693487 +0000 UTC m=+123.918198164" watchObservedRunningTime="2025-12-09 09:49:27.585329298 +0000 UTC m=+123.919833965" Dec 09 09:49:27 crc kubenswrapper[4824]: I1209 09:49:27.718352 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-5tn52"] Dec 09 09:49:27 crc kubenswrapper[4824]: I1209 09:49:27.718490 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:49:27 crc kubenswrapper[4824]: E1209 09:49:27.718606 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:49:28 crc kubenswrapper[4824]: I1209 09:49:28.910022 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:49:28 crc kubenswrapper[4824]: I1209 09:49:28.910080 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:49:28 crc kubenswrapper[4824]: I1209 09:49:28.910103 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:49:28 crc kubenswrapper[4824]: I1209 09:49:28.910022 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:49:28 crc kubenswrapper[4824]: E1209 09:49:28.910254 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:49:28 crc kubenswrapper[4824]: E1209 09:49:28.910483 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:49:28 crc kubenswrapper[4824]: E1209 09:49:28.910620 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:49:28 crc kubenswrapper[4824]: E1209 09:49:28.910749 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:49:29 crc kubenswrapper[4824]: E1209 09:49:29.101558 4824 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 09:49:30 crc kubenswrapper[4824]: I1209 09:49:30.909772 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:49:30 crc kubenswrapper[4824]: I1209 09:49:30.909891 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:49:30 crc kubenswrapper[4824]: E1209 09:49:30.910335 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:49:30 crc kubenswrapper[4824]: I1209 09:49:30.909961 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:49:30 crc kubenswrapper[4824]: E1209 09:49:30.910517 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:49:30 crc kubenswrapper[4824]: I1209 09:49:30.910660 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:49:30 crc kubenswrapper[4824]: E1209 09:49:30.910719 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:49:30 crc kubenswrapper[4824]: E1209 09:49:30.910774 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:49:30 crc kubenswrapper[4824]: I1209 09:49:30.911138 4824 scope.go:117] "RemoveContainer" containerID="c0ee66e89363676597efd13d02245ce8c9a9af288c04f2ed9b7ae8bed31da259" Dec 09 09:49:31 crc kubenswrapper[4824]: I1209 09:49:31.570086 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-pvksg_74aa3cd7-0abc-4fc9-9697-e23b693a7611/kube-multus/1.log" Dec 09 09:49:31 crc kubenswrapper[4824]: I1209 09:49:31.570152 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pvksg" event={"ID":"74aa3cd7-0abc-4fc9-9697-e23b693a7611","Type":"ContainerStarted","Data":"bf4f9aab454720c3762d743e8301467299a78c44fb18b35f40c349f4d60497d7"} Dec 09 09:49:32 crc kubenswrapper[4824]: I1209 09:49:32.909945 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:49:32 crc kubenswrapper[4824]: E1209 09:49:32.910104 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 09:49:32 crc kubenswrapper[4824]: I1209 09:49:32.910196 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:49:32 crc kubenswrapper[4824]: I1209 09:49:32.910228 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:49:32 crc kubenswrapper[4824]: I1209 09:49:32.910393 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:49:32 crc kubenswrapper[4824]: E1209 09:49:32.910397 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 09:49:32 crc kubenswrapper[4824]: E1209 09:49:32.910603 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 09:49:32 crc kubenswrapper[4824]: E1209 09:49:32.910688 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5tn52" podUID="fcbce9ea-3d55-4e09-b5bc-d5bb8c472219" Dec 09 09:49:34 crc kubenswrapper[4824]: I1209 09:49:34.909584 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:49:34 crc kubenswrapper[4824]: I1209 09:49:34.909727 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:49:34 crc kubenswrapper[4824]: I1209 09:49:34.909766 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:49:34 crc kubenswrapper[4824]: I1209 09:49:34.909811 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:49:34 crc kubenswrapper[4824]: I1209 09:49:34.912315 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 09 09:49:34 crc kubenswrapper[4824]: I1209 09:49:34.912643 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 09 09:49:34 crc kubenswrapper[4824]: I1209 09:49:34.913065 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 09 09:49:34 crc kubenswrapper[4824]: I1209 09:49:34.913310 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 09 09:49:34 crc kubenswrapper[4824]: I1209 09:49:34.912833 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 09 09:49:34 crc kubenswrapper[4824]: I1209 09:49:34.915036 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.457339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.493590 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.494047 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.497622 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.498466 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.498613 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.498861 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.499123 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.504218 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.506250 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-l5ld8"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.506675 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.507040 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qr8wb"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.507588 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qr8wb" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.509116 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zv6dz"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.509393 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zv6dz" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.510232 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tqhnj"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.510383 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.510426 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.510612 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.510665 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.510867 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.510390 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.511735 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-wgtx2"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.511796 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.512220 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.512950 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-bxplc"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.513295 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.513622 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.514171 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.520622 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.521276 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.521311 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.521321 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.521483 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.521473 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.521886 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.522337 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.522620 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.522624 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.522898 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.522828 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.522834 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.522854 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.525113 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.525257 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.526011 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.526132 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.526237 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.527107 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.527305 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.527421 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.527622 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.535746 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.536204 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.536591 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.536952 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.537496 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.537578 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.527831 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.538349 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.539232 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.540913 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.541471 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-847gg"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.544303 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-847gg" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.560801 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.561128 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.561496 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.562237 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.562552 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.565577 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-mjsbv"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.569267 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.570033 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.566172 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.565930 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.570807 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f93c02c2-b12a-489d-9f82-1ca3dadd18d5-serving-cert\") pod \"apiserver-7bbb656c7d-n8zgd\" (UID: \"f93c02c2-b12a-489d-9f82-1ca3dadd18d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.570925 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f93c02c2-b12a-489d-9f82-1ca3dadd18d5-audit-policies\") pod \"apiserver-7bbb656c7d-n8zgd\" (UID: \"f93c02c2-b12a-489d-9f82-1ca3dadd18d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.571052 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b18b84da-1277-4a24-b434-c93eb83a3af8-serving-cert\") pod \"controller-manager-879f6c89f-tqhnj\" (UID: \"b18b84da-1277-4a24-b434-c93eb83a3af8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.571161 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b18b84da-1277-4a24-b434-c93eb83a3af8-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-tqhnj\" (UID: \"b18b84da-1277-4a24-b434-c93eb83a3af8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.567096 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.571426 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-b9bnh"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.569588 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjsbv" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.571270 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.571848 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f93c02c2-b12a-489d-9f82-1ca3dadd18d5-encryption-config\") pod \"apiserver-7bbb656c7d-n8zgd\" (UID: \"f93c02c2-b12a-489d-9f82-1ca3dadd18d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.571897 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.571940 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b-serving-cert\") pod \"route-controller-manager-6576b87f9c-rqnfv\" (UID: \"2d573cb8-34fa-40ac-98e6-c4f18caf4e0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.571967 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572004 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a7eb77ab-add0-4601-a279-3f8fc956b52f-audit-policies\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572039 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b18b84da-1277-4a24-b434-c93eb83a3af8-client-ca\") pod \"controller-manager-879f6c89f-tqhnj\" (UID: \"b18b84da-1277-4a24-b434-c93eb83a3af8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572062 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smn5q\" (UniqueName: \"kubernetes.io/projected/c3400b7c-5666-4b36-a50f-7800f6737527-kube-api-access-smn5q\") pod \"authentication-operator-69f744f599-bxplc\" (UID: \"c3400b7c-5666-4b36-a50f-7800f6737527\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572089 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76dddf3f-2b2c-4644-ab68-402fa1e6b4d3-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zv6dz\" (UID: \"76dddf3f-2b2c-4644-ab68-402fa1e6b4d3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zv6dz" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572130 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b-client-ca\") pod \"route-controller-manager-6576b87f9c-rqnfv\" (UID: \"2d573cb8-34fa-40ac-98e6-c4f18caf4e0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572165 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.567391 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572188 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3400b7c-5666-4b36-a50f-7800f6737527-serving-cert\") pod \"authentication-operator-69f744f599-bxplc\" (UID: \"c3400b7c-5666-4b36-a50f-7800f6737527\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572300 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd6c14a0-99d8-485d-81f3-cc4ebe5a943b-serving-cert\") pod \"console-operator-58897d9998-wgtx2\" (UID: \"cd6c14a0-99d8-485d-81f3-cc4ebe5a943b\") " pod="openshift-console-operator/console-operator-58897d9998-wgtx2" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572326 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cd6c14a0-99d8-485d-81f3-cc4ebe5a943b-trusted-ca\") pod \"console-operator-58897d9998-wgtx2\" (UID: \"cd6c14a0-99d8-485d-81f3-cc4ebe5a943b\") " pod="openshift-console-operator/console-operator-58897d9998-wgtx2" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572356 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b-config\") pod \"route-controller-manager-6576b87f9c-rqnfv\" (UID: \"2d573cb8-34fa-40ac-98e6-c4f18caf4e0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572380 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76dddf3f-2b2c-4644-ab68-402fa1e6b4d3-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zv6dz\" (UID: \"76dddf3f-2b2c-4644-ab68-402fa1e6b4d3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zv6dz" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572418 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3400b7c-5666-4b36-a50f-7800f6737527-service-ca-bundle\") pod \"authentication-operator-69f744f599-bxplc\" (UID: \"c3400b7c-5666-4b36-a50f-7800f6737527\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572458 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572483 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f93c02c2-b12a-489d-9f82-1ca3dadd18d5-etcd-client\") pod \"apiserver-7bbb656c7d-n8zgd\" (UID: \"f93c02c2-b12a-489d-9f82-1ca3dadd18d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572532 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gh4vw\" (UniqueName: \"kubernetes.io/projected/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b-kube-api-access-gh4vw\") pod \"route-controller-manager-6576b87f9c-rqnfv\" (UID: \"2d573cb8-34fa-40ac-98e6-c4f18caf4e0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572557 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a7eb77ab-add0-4601-a279-3f8fc956b52f-audit-dir\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572580 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572600 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f93c02c2-b12a-489d-9f82-1ca3dadd18d5-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-n8zgd\" (UID: \"f93c02c2-b12a-489d-9f82-1ca3dadd18d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572624 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/66cd34f5-c1e9-4b01-80ba-71b574b97ced-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qr8wb\" (UID: \"66cd34f5-c1e9-4b01-80ba-71b574b97ced\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qr8wb" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572648 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd6c14a0-99d8-485d-81f3-cc4ebe5a943b-config\") pod \"console-operator-58897d9998-wgtx2\" (UID: \"cd6c14a0-99d8-485d-81f3-cc4ebe5a943b\") " pod="openshift-console-operator/console-operator-58897d9998-wgtx2" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572723 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w9xq\" (UniqueName: \"kubernetes.io/projected/66cd34f5-c1e9-4b01-80ba-71b574b97ced-kube-api-access-2w9xq\") pod \"cluster-samples-operator-665b6dd947-qr8wb\" (UID: \"66cd34f5-c1e9-4b01-80ba-71b574b97ced\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qr8wb" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572773 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9d9v8\" (UniqueName: \"kubernetes.io/projected/76dddf3f-2b2c-4644-ab68-402fa1e6b4d3-kube-api-access-9d9v8\") pod \"openshift-apiserver-operator-796bbdcf4f-zv6dz\" (UID: \"76dddf3f-2b2c-4644-ab68-402fa1e6b4d3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zv6dz" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572832 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh8zh\" (UniqueName: \"kubernetes.io/projected/b18b84da-1277-4a24-b434-c93eb83a3af8-kube-api-access-jh8zh\") pod \"controller-manager-879f6c89f-tqhnj\" (UID: \"b18b84da-1277-4a24-b434-c93eb83a3af8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572857 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3400b7c-5666-4b36-a50f-7800f6737527-config\") pod \"authentication-operator-69f744f599-bxplc\" (UID: \"c3400b7c-5666-4b36-a50f-7800f6737527\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572881 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3400b7c-5666-4b36-a50f-7800f6737527-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-bxplc\" (UID: \"c3400b7c-5666-4b36-a50f-7800f6737527\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572928 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572953 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572974 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f93c02c2-b12a-489d-9f82-1ca3dadd18d5-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-n8zgd\" (UID: \"f93c02c2-b12a-489d-9f82-1ca3dadd18d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.572996 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.573022 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b18b84da-1277-4a24-b434-c93eb83a3af8-config\") pod \"controller-manager-879f6c89f-tqhnj\" (UID: \"b18b84da-1277-4a24-b434-c93eb83a3af8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.573050 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7bfh\" (UniqueName: \"kubernetes.io/projected/f93c02c2-b12a-489d-9f82-1ca3dadd18d5-kube-api-access-x7bfh\") pod \"apiserver-7bbb656c7d-n8zgd\" (UID: \"f93c02c2-b12a-489d-9f82-1ca3dadd18d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.573070 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6thw8\" (UniqueName: \"kubernetes.io/projected/cd6c14a0-99d8-485d-81f3-cc4ebe5a943b-kube-api-access-6thw8\") pod \"console-operator-58897d9998-wgtx2\" (UID: \"cd6c14a0-99d8-485d-81f3-cc4ebe5a943b\") " pod="openshift-console-operator/console-operator-58897d9998-wgtx2" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.573103 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.573131 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.573157 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2zqr\" (UniqueName: \"kubernetes.io/projected/a7eb77ab-add0-4601-a279-3f8fc956b52f-kube-api-access-v2zqr\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.573179 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f93c02c2-b12a-489d-9f82-1ca3dadd18d5-audit-dir\") pod \"apiserver-7bbb656c7d-n8zgd\" (UID: \"f93c02c2-b12a-489d-9f82-1ca3dadd18d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.573774 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qkkdx"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.574213 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4rv7m"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.574828 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7b9dd"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.576454 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-7b9dd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.575019 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.575722 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-4rv7m" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.579257 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.582256 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.582673 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-c6559"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.583492 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-c6559" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.586201 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.587173 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.588414 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pd82d"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.589111 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-ldw7r"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.589446 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.589765 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.594336 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-b9bnh" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.601485 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qr8wb"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.603534 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.603931 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.607740 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.608044 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.608226 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.608729 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.609208 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.609303 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.609524 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.609590 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-l5ld8"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.609613 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.609630 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-jgwrj"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.610033 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.610060 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.610150 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.610234 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.610287 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.610299 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.610518 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.610594 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.610677 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.610875 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.611214 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.611256 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.611482 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.612718 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.612797 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.613640 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.613761 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.614167 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.618867 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.631492 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jgwrj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.640150 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2t6n7"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.640936 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-k4fpg"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.641472 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x7j5h"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.643137 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x7j5h" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.643150 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k4fpg" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.641911 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2t6n7" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.634259 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.644292 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pvgjx"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.635956 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.645903 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmk4v"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.636058 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.637085 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.646289 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmk4v" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.646457 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pvgjx" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.637128 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.641822 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.642102 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.642261 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.642295 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.642323 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.642360 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.642543 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.642580 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.644929 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.645105 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.653476 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-mj7jx"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.654090 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.657452 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q8c9m"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.658076 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q8c9m" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.659771 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-zl4mr"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.660115 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.660461 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zl4mr" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.660596 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.662409 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421225-96fgb"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.663045 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421225-96fgb" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.663354 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kv8sn"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.663681 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kv8sn" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.664447 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-g2cbw"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.665057 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-g2cbw" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.665762 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9jtv8"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.666612 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9jtv8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.666822 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-99q6g"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.668331 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-99q6g" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.669014 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.670084 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.670402 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-hjg74"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.671358 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-bxplc"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.671639 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hjg74" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.672271 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.673806 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.673999 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.675045 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-pk58m"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.675539 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-pk58m" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.676402 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f93c02c2-b12a-489d-9f82-1ca3dadd18d5-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-n8zgd\" (UID: \"f93c02c2-b12a-489d-9f82-1ca3dadd18d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.676440 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dpmx\" (UniqueName: \"kubernetes.io/projected/00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53-kube-api-access-4dpmx\") pod \"ingress-operator-5b745b69d9-jgwrj\" (UID: \"00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jgwrj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.676474 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.676497 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.676523 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.676548 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b18b84da-1277-4a24-b434-c93eb83a3af8-config\") pod \"controller-manager-879f6c89f-tqhnj\" (UID: \"b18b84da-1277-4a24-b434-c93eb83a3af8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.676569 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7bfh\" (UniqueName: \"kubernetes.io/projected/f93c02c2-b12a-489d-9f82-1ca3dadd18d5-kube-api-access-x7bfh\") pod \"apiserver-7bbb656c7d-n8zgd\" (UID: \"f93c02c2-b12a-489d-9f82-1ca3dadd18d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.676592 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f4616db0-42ba-4a3b-88f2-57a0663d0a15-etcd-ca\") pod \"etcd-operator-b45778765-7b9dd\" (UID: \"f4616db0-42ba-4a3b-88f2-57a0663d0a15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b9dd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.676625 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6thw8\" (UniqueName: \"kubernetes.io/projected/cd6c14a0-99d8-485d-81f3-cc4ebe5a943b-kube-api-access-6thw8\") pod \"console-operator-58897d9998-wgtx2\" (UID: \"cd6c14a0-99d8-485d-81f3-cc4ebe5a943b\") " pod="openshift-console-operator/console-operator-58897d9998-wgtx2" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.676649 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.676704 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.677020 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f93c02c2-b12a-489d-9f82-1ca3dadd18d5-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-n8zgd\" (UID: \"f93c02c2-b12a-489d-9f82-1ca3dadd18d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.676733 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2zqr\" (UniqueName: \"kubernetes.io/projected/a7eb77ab-add0-4601-a279-3f8fc956b52f-kube-api-access-v2zqr\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.677302 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f93c02c2-b12a-489d-9f82-1ca3dadd18d5-audit-dir\") pod \"apiserver-7bbb656c7d-n8zgd\" (UID: \"f93c02c2-b12a-489d-9f82-1ca3dadd18d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.677362 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtwxj\" (UniqueName: \"kubernetes.io/projected/76df4be1-7bf3-461d-a9d7-6a2be7c0e425-kube-api-access-vtwxj\") pod \"machine-config-operator-74547568cd-k4fpg\" (UID: \"76df4be1-7bf3-461d-a9d7-6a2be7c0e425\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k4fpg" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.677419 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.677448 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f93c02c2-b12a-489d-9f82-1ca3dadd18d5-serving-cert\") pod \"apiserver-7bbb656c7d-n8zgd\" (UID: \"f93c02c2-b12a-489d-9f82-1ca3dadd18d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.677470 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4616db0-42ba-4a3b-88f2-57a0663d0a15-serving-cert\") pod \"etcd-operator-b45778765-7b9dd\" (UID: \"f4616db0-42ba-4a3b-88f2-57a0663d0a15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b9dd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.677493 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f93c02c2-b12a-489d-9f82-1ca3dadd18d5-audit-policies\") pod \"apiserver-7bbb656c7d-n8zgd\" (UID: \"f93c02c2-b12a-489d-9f82-1ca3dadd18d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.677514 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b18b84da-1277-4a24-b434-c93eb83a3af8-serving-cert\") pod \"controller-manager-879f6c89f-tqhnj\" (UID: \"b18b84da-1277-4a24-b434-c93eb83a3af8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.677540 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b18b84da-1277-4a24-b434-c93eb83a3af8-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-tqhnj\" (UID: \"b18b84da-1277-4a24-b434-c93eb83a3af8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.677569 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.677593 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f93c02c2-b12a-489d-9f82-1ca3dadd18d5-encryption-config\") pod \"apiserver-7bbb656c7d-n8zgd\" (UID: \"f93c02c2-b12a-489d-9f82-1ca3dadd18d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.677616 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f4616db0-42ba-4a3b-88f2-57a0663d0a15-etcd-service-ca\") pod \"etcd-operator-b45778765-7b9dd\" (UID: \"f4616db0-42ba-4a3b-88f2-57a0663d0a15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b9dd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.677645 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.677670 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b-serving-cert\") pod \"route-controller-manager-6576b87f9c-rqnfv\" (UID: \"2d573cb8-34fa-40ac-98e6-c4f18caf4e0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.677696 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b18b84da-1277-4a24-b434-c93eb83a3af8-client-ca\") pod \"controller-manager-879f6c89f-tqhnj\" (UID: \"b18b84da-1277-4a24-b434-c93eb83a3af8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.677722 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smn5q\" (UniqueName: \"kubernetes.io/projected/c3400b7c-5666-4b36-a50f-7800f6737527-kube-api-access-smn5q\") pod \"authentication-operator-69f744f599-bxplc\" (UID: \"c3400b7c-5666-4b36-a50f-7800f6737527\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.677760 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a7eb77ab-add0-4601-a279-3f8fc956b52f-audit-policies\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.677800 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5047f85e-0349-40fc-b966-1956bd66cbbc-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-x7j5h\" (UID: \"5047f85e-0349-40fc-b966-1956bd66cbbc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x7j5h" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.677829 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0c5e7133-21dc-46e0-b334-9d36c0427aa3-trusted-ca-bundle\") pod \"console-f9d7485db-ldw7r\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.677858 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76dddf3f-2b2c-4644-ab68-402fa1e6b4d3-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zv6dz\" (UID: \"76dddf3f-2b2c-4644-ab68-402fa1e6b4d3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zv6dz" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.677881 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/76df4be1-7bf3-461d-a9d7-6a2be7c0e425-images\") pod \"machine-config-operator-74547568cd-k4fpg\" (UID: \"76df4be1-7bf3-461d-a9d7-6a2be7c0e425\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k4fpg" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.677911 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0c5e7133-21dc-46e0-b334-9d36c0427aa3-console-config\") pod \"console-f9d7485db-ldw7r\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.677935 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b-client-ca\") pod \"route-controller-manager-6576b87f9c-rqnfv\" (UID: \"2d573cb8-34fa-40ac-98e6-c4f18caf4e0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.677982 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678022 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3400b7c-5666-4b36-a50f-7800f6737527-serving-cert\") pod \"authentication-operator-69f744f599-bxplc\" (UID: \"c3400b7c-5666-4b36-a50f-7800f6737527\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678069 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/76df4be1-7bf3-461d-a9d7-6a2be7c0e425-auth-proxy-config\") pod \"machine-config-operator-74547568cd-k4fpg\" (UID: \"76df4be1-7bf3-461d-a9d7-6a2be7c0e425\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k4fpg" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678093 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0c5e7133-21dc-46e0-b334-9d36c0427aa3-service-ca\") pod \"console-f9d7485db-ldw7r\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678120 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd6c14a0-99d8-485d-81f3-cc4ebe5a943b-serving-cert\") pod \"console-operator-58897d9998-wgtx2\" (UID: \"cd6c14a0-99d8-485d-81f3-cc4ebe5a943b\") " pod="openshift-console-operator/console-operator-58897d9998-wgtx2" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678143 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53-bound-sa-token\") pod \"ingress-operator-5b745b69d9-jgwrj\" (UID: \"00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jgwrj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678167 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cd6c14a0-99d8-485d-81f3-cc4ebe5a943b-trusted-ca\") pod \"console-operator-58897d9998-wgtx2\" (UID: \"cd6c14a0-99d8-485d-81f3-cc4ebe5a943b\") " pod="openshift-console-operator/console-operator-58897d9998-wgtx2" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678193 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b-config\") pod \"route-controller-manager-6576b87f9c-rqnfv\" (UID: \"2d573cb8-34fa-40ac-98e6-c4f18caf4e0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678214 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6042da41-13ef-460e-aee4-491c7758f635-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2t6n7\" (UID: \"6042da41-13ef-460e-aee4-491c7758f635\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2t6n7" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678234 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t985r\" (UniqueName: \"kubernetes.io/projected/5047f85e-0349-40fc-b966-1956bd66cbbc-kube-api-access-t985r\") pod \"kube-storage-version-migrator-operator-b67b599dd-x7j5h\" (UID: \"5047f85e-0349-40fc-b966-1956bd66cbbc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x7j5h" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678243 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678260 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76dddf3f-2b2c-4644-ab68-402fa1e6b4d3-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zv6dz\" (UID: \"76dddf3f-2b2c-4644-ab68-402fa1e6b4d3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zv6dz" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678284 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3400b7c-5666-4b36-a50f-7800f6737527-service-ca-bundle\") pod \"authentication-operator-69f744f599-bxplc\" (UID: \"c3400b7c-5666-4b36-a50f-7800f6737527\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678308 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53-metrics-tls\") pod \"ingress-operator-5b745b69d9-jgwrj\" (UID: \"00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jgwrj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678329 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0c5e7133-21dc-46e0-b334-9d36c0427aa3-oauth-serving-cert\") pod \"console-f9d7485db-ldw7r\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678357 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678378 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f93c02c2-b12a-489d-9f82-1ca3dadd18d5-etcd-client\") pod \"apiserver-7bbb656c7d-n8zgd\" (UID: \"f93c02c2-b12a-489d-9f82-1ca3dadd18d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678402 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0c5e7133-21dc-46e0-b334-9d36c0427aa3-console-serving-cert\") pod \"console-f9d7485db-ldw7r\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678425 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53-trusted-ca\") pod \"ingress-operator-5b745b69d9-jgwrj\" (UID: \"00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jgwrj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678449 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/76df4be1-7bf3-461d-a9d7-6a2be7c0e425-proxy-tls\") pod \"machine-config-operator-74547568cd-k4fpg\" (UID: \"76df4be1-7bf3-461d-a9d7-6a2be7c0e425\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k4fpg" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678480 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gh4vw\" (UniqueName: \"kubernetes.io/projected/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b-kube-api-access-gh4vw\") pod \"route-controller-manager-6576b87f9c-rqnfv\" (UID: \"2d573cb8-34fa-40ac-98e6-c4f18caf4e0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678501 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b18b84da-1277-4a24-b434-c93eb83a3af8-config\") pod \"controller-manager-879f6c89f-tqhnj\" (UID: \"b18b84da-1277-4a24-b434-c93eb83a3af8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678505 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6km2\" (UniqueName: \"kubernetes.io/projected/f4616db0-42ba-4a3b-88f2-57a0663d0a15-kube-api-access-k6km2\") pod \"etcd-operator-b45778765-7b9dd\" (UID: \"f4616db0-42ba-4a3b-88f2-57a0663d0a15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b9dd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678555 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6042da41-13ef-460e-aee4-491c7758f635-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2t6n7\" (UID: \"6042da41-13ef-460e-aee4-491c7758f635\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2t6n7" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678582 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6042da41-13ef-460e-aee4-491c7758f635-config\") pod \"kube-apiserver-operator-766d6c64bb-2t6n7\" (UID: \"6042da41-13ef-460e-aee4-491c7758f635\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2t6n7" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678619 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a7eb77ab-add0-4601-a279-3f8fc956b52f-audit-dir\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678645 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678670 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f93c02c2-b12a-489d-9f82-1ca3dadd18d5-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-n8zgd\" (UID: \"f93c02c2-b12a-489d-9f82-1ca3dadd18d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678693 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/66cd34f5-c1e9-4b01-80ba-71b574b97ced-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qr8wb\" (UID: \"66cd34f5-c1e9-4b01-80ba-71b574b97ced\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qr8wb" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678716 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd6c14a0-99d8-485d-81f3-cc4ebe5a943b-config\") pod \"console-operator-58897d9998-wgtx2\" (UID: \"cd6c14a0-99d8-485d-81f3-cc4ebe5a943b\") " pod="openshift-console-operator/console-operator-58897d9998-wgtx2" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678738 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w9xq\" (UniqueName: \"kubernetes.io/projected/66cd34f5-c1e9-4b01-80ba-71b574b97ced-kube-api-access-2w9xq\") pod \"cluster-samples-operator-665b6dd947-qr8wb\" (UID: \"66cd34f5-c1e9-4b01-80ba-71b574b97ced\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qr8wb" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678761 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f4616db0-42ba-4a3b-88f2-57a0663d0a15-etcd-client\") pod \"etcd-operator-b45778765-7b9dd\" (UID: \"f4616db0-42ba-4a3b-88f2-57a0663d0a15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b9dd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678804 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdcvb\" (UniqueName: \"kubernetes.io/projected/0c5e7133-21dc-46e0-b334-9d36c0427aa3-kube-api-access-zdcvb\") pod \"console-f9d7485db-ldw7r\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678847 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9d9v8\" (UniqueName: \"kubernetes.io/projected/76dddf3f-2b2c-4644-ab68-402fa1e6b4d3-kube-api-access-9d9v8\") pod \"openshift-apiserver-operator-796bbdcf4f-zv6dz\" (UID: \"76dddf3f-2b2c-4644-ab68-402fa1e6b4d3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zv6dz" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678875 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5047f85e-0349-40fc-b966-1956bd66cbbc-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-x7j5h\" (UID: \"5047f85e-0349-40fc-b966-1956bd66cbbc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x7j5h" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678904 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3400b7c-5666-4b36-a50f-7800f6737527-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-bxplc\" (UID: \"c3400b7c-5666-4b36-a50f-7800f6737527\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678925 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4616db0-42ba-4a3b-88f2-57a0663d0a15-config\") pod \"etcd-operator-b45778765-7b9dd\" (UID: \"f4616db0-42ba-4a3b-88f2-57a0663d0a15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b9dd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678950 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh8zh\" (UniqueName: \"kubernetes.io/projected/b18b84da-1277-4a24-b434-c93eb83a3af8-kube-api-access-jh8zh\") pod \"controller-manager-879f6c89f-tqhnj\" (UID: \"b18b84da-1277-4a24-b434-c93eb83a3af8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.678973 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3400b7c-5666-4b36-a50f-7800f6737527-config\") pod \"authentication-operator-69f744f599-bxplc\" (UID: \"c3400b7c-5666-4b36-a50f-7800f6737527\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.679004 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0c5e7133-21dc-46e0-b334-9d36c0427aa3-console-oauth-config\") pod \"console-f9d7485db-ldw7r\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.679047 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.679145 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a7eb77ab-add0-4601-a279-3f8fc956b52f-audit-dir\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.679584 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b-client-ca\") pod \"route-controller-manager-6576b87f9c-rqnfv\" (UID: \"2d573cb8-34fa-40ac-98e6-c4f18caf4e0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.679766 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f93c02c2-b12a-489d-9f82-1ca3dadd18d5-audit-dir\") pod \"apiserver-7bbb656c7d-n8zgd\" (UID: \"f93c02c2-b12a-489d-9f82-1ca3dadd18d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.679862 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3400b7c-5666-4b36-a50f-7800f6737527-service-ca-bundle\") pod \"authentication-operator-69f744f599-bxplc\" (UID: \"c3400b7c-5666-4b36-a50f-7800f6737527\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.680394 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f93c02c2-b12a-489d-9f82-1ca3dadd18d5-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-n8zgd\" (UID: \"f93c02c2-b12a-489d-9f82-1ca3dadd18d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.681139 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3400b7c-5666-4b36-a50f-7800f6737527-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-bxplc\" (UID: \"c3400b7c-5666-4b36-a50f-7800f6737527\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.681644 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f93c02c2-b12a-489d-9f82-1ca3dadd18d5-audit-policies\") pod \"apiserver-7bbb656c7d-n8zgd\" (UID: \"f93c02c2-b12a-489d-9f82-1ca3dadd18d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.699159 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b18b84da-1277-4a24-b434-c93eb83a3af8-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-tqhnj\" (UID: \"b18b84da-1277-4a24-b434-c93eb83a3af8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.700620 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.700842 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd6c14a0-99d8-485d-81f3-cc4ebe5a943b-config\") pod \"console-operator-58897d9998-wgtx2\" (UID: \"cd6c14a0-99d8-485d-81f3-cc4ebe5a943b\") " pod="openshift-console-operator/console-operator-58897d9998-wgtx2" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.701262 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f93c02c2-b12a-489d-9f82-1ca3dadd18d5-serving-cert\") pod \"apiserver-7bbb656c7d-n8zgd\" (UID: \"f93c02c2-b12a-489d-9f82-1ca3dadd18d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.702230 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/66cd34f5-c1e9-4b01-80ba-71b574b97ced-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qr8wb\" (UID: \"66cd34f5-c1e9-4b01-80ba-71b574b97ced\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qr8wb" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.702898 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cd6c14a0-99d8-485d-81f3-cc4ebe5a943b-trusted-ca\") pod \"console-operator-58897d9998-wgtx2\" (UID: \"cd6c14a0-99d8-485d-81f3-cc4ebe5a943b\") " pod="openshift-console-operator/console-operator-58897d9998-wgtx2" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.703484 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b-config\") pod \"route-controller-manager-6576b87f9c-rqnfv\" (UID: \"2d573cb8-34fa-40ac-98e6-c4f18caf4e0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.704351 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pd82d"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.704495 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.705082 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd6c14a0-99d8-485d-81f3-cc4ebe5a943b-serving-cert\") pod \"console-operator-58897d9998-wgtx2\" (UID: \"cd6c14a0-99d8-485d-81f3-cc4ebe5a943b\") " pod="openshift-console-operator/console-operator-58897d9998-wgtx2" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.707510 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.708290 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76dddf3f-2b2c-4644-ab68-402fa1e6b4d3-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zv6dz\" (UID: \"76dddf3f-2b2c-4644-ab68-402fa1e6b4d3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zv6dz" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.708524 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b18b84da-1277-4a24-b434-c93eb83a3af8-client-ca\") pod \"controller-manager-879f6c89f-tqhnj\" (UID: \"b18b84da-1277-4a24-b434-c93eb83a3af8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.679148 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.703895 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.708461 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3400b7c-5666-4b36-a50f-7800f6737527-config\") pod \"authentication-operator-69f744f599-bxplc\" (UID: \"c3400b7c-5666-4b36-a50f-7800f6737527\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.710376 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a7eb77ab-add0-4601-a279-3f8fc956b52f-audit-policies\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.710389 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76dddf3f-2b2c-4644-ab68-402fa1e6b4d3-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zv6dz\" (UID: \"76dddf3f-2b2c-4644-ab68-402fa1e6b4d3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zv6dz" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.710460 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b18b84da-1277-4a24-b434-c93eb83a3af8-serving-cert\") pod \"controller-manager-879f6c89f-tqhnj\" (UID: \"b18b84da-1277-4a24-b434-c93eb83a3af8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.710506 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.710665 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f93c02c2-b12a-489d-9f82-1ca3dadd18d5-etcd-client\") pod \"apiserver-7bbb656c7d-n8zgd\" (UID: \"f93c02c2-b12a-489d-9f82-1ca3dadd18d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.707546 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.715446 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b-serving-cert\") pod \"route-controller-manager-6576b87f9c-rqnfv\" (UID: \"2d573cb8-34fa-40ac-98e6-c4f18caf4e0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.716731 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.716944 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.718366 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.718429 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c2dfm"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.720529 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3400b7c-5666-4b36-a50f-7800f6737527-serving-cert\") pod \"authentication-operator-69f744f599-bxplc\" (UID: \"c3400b7c-5666-4b36-a50f-7800f6737527\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.722090 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.723030 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.724348 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-c6559"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.724552 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c2dfm" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.725377 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f93c02c2-b12a-489d-9f82-1ca3dadd18d5-encryption-config\") pod \"apiserver-7bbb656c7d-n8zgd\" (UID: \"f93c02c2-b12a-489d-9f82-1ca3dadd18d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.728254 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-wgtx2"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.730360 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.731360 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zv6dz"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.732541 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.734511 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-847gg"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.736598 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.739455 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-k4fpg"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.740905 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2t6n7"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.742547 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-jgwrj"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.744426 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-zl4mr"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.745717 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pvgjx"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.747022 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-mnksj"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.748491 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-b9bnh"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.748583 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-mnksj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.751019 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x7j5h"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.754753 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9jtv8"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.757010 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.758116 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.759729 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.761052 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.763452 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmk4v"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.764881 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tqhnj"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.765918 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-ldw7r"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.767588 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qkkdx"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.768615 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7b9dd"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.770091 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kv8sn"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.771066 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-c9nzk"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.771727 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-c9nzk" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.772148 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q8c9m"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.774152 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421225-96fgb"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.776177 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.776408 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.777826 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.779369 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-pk58m"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.779900 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/76df4be1-7bf3-461d-a9d7-6a2be7c0e425-images\") pod \"machine-config-operator-74547568cd-k4fpg\" (UID: \"76df4be1-7bf3-461d-a9d7-6a2be7c0e425\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k4fpg" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.779934 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0c5e7133-21dc-46e0-b334-9d36c0427aa3-console-config\") pod \"console-f9d7485db-ldw7r\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.779973 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/76df4be1-7bf3-461d-a9d7-6a2be7c0e425-auth-proxy-config\") pod \"machine-config-operator-74547568cd-k4fpg\" (UID: \"76df4be1-7bf3-461d-a9d7-6a2be7c0e425\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k4fpg" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780025 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0c5e7133-21dc-46e0-b334-9d36c0427aa3-service-ca\") pod \"console-f9d7485db-ldw7r\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780048 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53-bound-sa-token\") pod \"ingress-operator-5b745b69d9-jgwrj\" (UID: \"00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jgwrj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780068 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6042da41-13ef-460e-aee4-491c7758f635-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2t6n7\" (UID: \"6042da41-13ef-460e-aee4-491c7758f635\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2t6n7" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780091 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53-metrics-tls\") pod \"ingress-operator-5b745b69d9-jgwrj\" (UID: \"00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jgwrj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780113 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t985r\" (UniqueName: \"kubernetes.io/projected/5047f85e-0349-40fc-b966-1956bd66cbbc-kube-api-access-t985r\") pod \"kube-storage-version-migrator-operator-b67b599dd-x7j5h\" (UID: \"5047f85e-0349-40fc-b966-1956bd66cbbc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x7j5h" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780137 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0c5e7133-21dc-46e0-b334-9d36c0427aa3-console-serving-cert\") pod \"console-f9d7485db-ldw7r\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780156 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0c5e7133-21dc-46e0-b334-9d36c0427aa3-oauth-serving-cert\") pod \"console-f9d7485db-ldw7r\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780189 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53-trusted-ca\") pod \"ingress-operator-5b745b69d9-jgwrj\" (UID: \"00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jgwrj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780210 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/76df4be1-7bf3-461d-a9d7-6a2be7c0e425-proxy-tls\") pod \"machine-config-operator-74547568cd-k4fpg\" (UID: \"76df4be1-7bf3-461d-a9d7-6a2be7c0e425\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k4fpg" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780230 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6km2\" (UniqueName: \"kubernetes.io/projected/f4616db0-42ba-4a3b-88f2-57a0663d0a15-kube-api-access-k6km2\") pod \"etcd-operator-b45778765-7b9dd\" (UID: \"f4616db0-42ba-4a3b-88f2-57a0663d0a15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b9dd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780252 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6042da41-13ef-460e-aee4-491c7758f635-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2t6n7\" (UID: \"6042da41-13ef-460e-aee4-491c7758f635\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2t6n7" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780272 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6042da41-13ef-460e-aee4-491c7758f635-config\") pod \"kube-apiserver-operator-766d6c64bb-2t6n7\" (UID: \"6042da41-13ef-460e-aee4-491c7758f635\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2t6n7" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780303 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f4616db0-42ba-4a3b-88f2-57a0663d0a15-etcd-client\") pod \"etcd-operator-b45778765-7b9dd\" (UID: \"f4616db0-42ba-4a3b-88f2-57a0663d0a15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b9dd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780325 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdcvb\" (UniqueName: \"kubernetes.io/projected/0c5e7133-21dc-46e0-b334-9d36c0427aa3-kube-api-access-zdcvb\") pod \"console-f9d7485db-ldw7r\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780368 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5047f85e-0349-40fc-b966-1956bd66cbbc-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-x7j5h\" (UID: \"5047f85e-0349-40fc-b966-1956bd66cbbc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x7j5h" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780396 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4616db0-42ba-4a3b-88f2-57a0663d0a15-config\") pod \"etcd-operator-b45778765-7b9dd\" (UID: \"f4616db0-42ba-4a3b-88f2-57a0663d0a15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b9dd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780416 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0c5e7133-21dc-46e0-b334-9d36c0427aa3-console-oauth-config\") pod \"console-f9d7485db-ldw7r\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780452 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dpmx\" (UniqueName: \"kubernetes.io/projected/00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53-kube-api-access-4dpmx\") pod \"ingress-operator-5b745b69d9-jgwrj\" (UID: \"00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jgwrj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780479 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f4616db0-42ba-4a3b-88f2-57a0663d0a15-etcd-ca\") pod \"etcd-operator-b45778765-7b9dd\" (UID: \"f4616db0-42ba-4a3b-88f2-57a0663d0a15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b9dd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780530 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtwxj\" (UniqueName: \"kubernetes.io/projected/76df4be1-7bf3-461d-a9d7-6a2be7c0e425-kube-api-access-vtwxj\") pod \"machine-config-operator-74547568cd-k4fpg\" (UID: \"76df4be1-7bf3-461d-a9d7-6a2be7c0e425\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k4fpg" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780550 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4616db0-42ba-4a3b-88f2-57a0663d0a15-serving-cert\") pod \"etcd-operator-b45778765-7b9dd\" (UID: \"f4616db0-42ba-4a3b-88f2-57a0663d0a15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b9dd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780576 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f4616db0-42ba-4a3b-88f2-57a0663d0a15-etcd-service-ca\") pod \"etcd-operator-b45778765-7b9dd\" (UID: \"f4616db0-42ba-4a3b-88f2-57a0663d0a15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b9dd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780695 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5047f85e-0349-40fc-b966-1956bd66cbbc-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-x7j5h\" (UID: \"5047f85e-0349-40fc-b966-1956bd66cbbc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x7j5h" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780719 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0c5e7133-21dc-46e0-b334-9d36c0427aa3-trusted-ca-bundle\") pod \"console-f9d7485db-ldw7r\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780718 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/76df4be1-7bf3-461d-a9d7-6a2be7c0e425-auth-proxy-config\") pod \"machine-config-operator-74547568cd-k4fpg\" (UID: \"76df4be1-7bf3-461d-a9d7-6a2be7c0e425\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k4fpg" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.780886 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4rv7m"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.781209 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0c5e7133-21dc-46e0-b334-9d36c0427aa3-console-config\") pod \"console-f9d7485db-ldw7r\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.782245 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f4616db0-42ba-4a3b-88f2-57a0663d0a15-etcd-service-ca\") pod \"etcd-operator-b45778765-7b9dd\" (UID: \"f4616db0-42ba-4a3b-88f2-57a0663d0a15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b9dd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.782579 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f4616db0-42ba-4a3b-88f2-57a0663d0a15-etcd-ca\") pod \"etcd-operator-b45778765-7b9dd\" (UID: \"f4616db0-42ba-4a3b-88f2-57a0663d0a15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b9dd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.782754 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c2dfm"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.782953 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4616db0-42ba-4a3b-88f2-57a0663d0a15-config\") pod \"etcd-operator-b45778765-7b9dd\" (UID: \"f4616db0-42ba-4a3b-88f2-57a0663d0a15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b9dd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.783484 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0c5e7133-21dc-46e0-b334-9d36c0427aa3-oauth-serving-cert\") pod \"console-f9d7485db-ldw7r\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.783543 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0c5e7133-21dc-46e0-b334-9d36c0427aa3-trusted-ca-bundle\") pod \"console-f9d7485db-ldw7r\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.785134 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4616db0-42ba-4a3b-88f2-57a0663d0a15-serving-cert\") pod \"etcd-operator-b45778765-7b9dd\" (UID: \"f4616db0-42ba-4a3b-88f2-57a0663d0a15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b9dd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.785175 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-99q6g"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.785427 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0c5e7133-21dc-46e0-b334-9d36c0427aa3-console-serving-cert\") pod \"console-f9d7485db-ldw7r\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.785484 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0c5e7133-21dc-46e0-b334-9d36c0427aa3-console-oauth-config\") pod \"console-f9d7485db-ldw7r\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.786128 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-c9nzk"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.787610 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-mnksj"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.787963 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f4616db0-42ba-4a3b-88f2-57a0663d0a15-etcd-client\") pod \"etcd-operator-b45778765-7b9dd\" (UID: \"f4616db0-42ba-4a3b-88f2-57a0663d0a15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b9dd" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.788890 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-hjg74"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.789907 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-zjjp6"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.791128 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-vsdpm"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.791581 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-zjjp6" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.794206 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0c5e7133-21dc-46e0-b334-9d36c0427aa3-service-ca\") pod \"console-f9d7485db-ldw7r\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.794470 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-zjjp6"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.794496 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-vsdpm"] Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.796277 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.796965 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.816731 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.836301 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.870567 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.878815 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.896401 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.905546 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53-metrics-tls\") pod \"ingress-operator-5b745b69d9-jgwrj\" (UID: \"00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jgwrj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.917181 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.944481 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.953272 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53-trusted-ca\") pod \"ingress-operator-5b745b69d9-jgwrj\" (UID: \"00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jgwrj" Dec 09 09:49:39 crc kubenswrapper[4824]: I1209 09:49:39.957123 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.017742 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.025944 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5047f85e-0349-40fc-b966-1956bd66cbbc-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-x7j5h\" (UID: \"5047f85e-0349-40fc-b966-1956bd66cbbc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x7j5h" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.036978 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.056479 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.077439 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.096972 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.116265 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.126235 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/76df4be1-7bf3-461d-a9d7-6a2be7c0e425-proxy-tls\") pod \"machine-config-operator-74547568cd-k4fpg\" (UID: \"76df4be1-7bf3-461d-a9d7-6a2be7c0e425\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k4fpg" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.136841 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.156330 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.165976 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6042da41-13ef-460e-aee4-491c7758f635-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2t6n7\" (UID: \"6042da41-13ef-460e-aee4-491c7758f635\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2t6n7" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.176910 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.182264 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6042da41-13ef-460e-aee4-491c7758f635-config\") pod \"kube-apiserver-operator-766d6c64bb-2t6n7\" (UID: \"6042da41-13ef-460e-aee4-491c7758f635\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2t6n7" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.196523 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.217042 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.237992 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.257040 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.277134 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.297501 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.316658 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.336757 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.356301 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.377318 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.396574 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.403262 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5047f85e-0349-40fc-b966-1956bd66cbbc-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-x7j5h\" (UID: \"5047f85e-0349-40fc-b966-1956bd66cbbc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x7j5h" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.415671 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.421900 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/76df4be1-7bf3-461d-a9d7-6a2be7c0e425-images\") pod \"machine-config-operator-74547568cd-k4fpg\" (UID: \"76df4be1-7bf3-461d-a9d7-6a2be7c0e425\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k4fpg" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.437263 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.457056 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.476741 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.496473 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.516846 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.537373 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.556896 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.576027 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.596919 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.616361 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.637459 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.655963 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.674915 4824 request.go:700] Waited for 1.014191968s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator/secrets?fieldSelector=metadata.name%3Dkube-storage-version-migrator-sa-dockercfg-5xfcg&limit=500&resourceVersion=0 Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.676752 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.698286 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.716583 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.736469 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.756505 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.776220 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.795874 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.815841 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.835797 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.856206 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.876646 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.911968 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.916192 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.936416 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.956862 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.976188 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 09 09:49:40 crc kubenswrapper[4824]: I1209 09:49:40.997174 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.016887 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.037470 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.056737 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.076370 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.096532 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.116839 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.142938 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.181277 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.185566 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.197010 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.216577 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.236814 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.256361 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.277594 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.297941 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.333835 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6thw8\" (UniqueName: \"kubernetes.io/projected/cd6c14a0-99d8-485d-81f3-cc4ebe5a943b-kube-api-access-6thw8\") pod \"console-operator-58897d9998-wgtx2\" (UID: \"cd6c14a0-99d8-485d-81f3-cc4ebe5a943b\") " pod="openshift-console-operator/console-operator-58897d9998-wgtx2" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.365483 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7bfh\" (UniqueName: \"kubernetes.io/projected/f93c02c2-b12a-489d-9f82-1ca3dadd18d5-kube-api-access-x7bfh\") pod \"apiserver-7bbb656c7d-n8zgd\" (UID: \"f93c02c2-b12a-489d-9f82-1ca3dadd18d5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.380333 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gh4vw\" (UniqueName: \"kubernetes.io/projected/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b-kube-api-access-gh4vw\") pod \"route-controller-manager-6576b87f9c-rqnfv\" (UID: \"2d573cb8-34fa-40ac-98e6-c4f18caf4e0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.402290 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2zqr\" (UniqueName: \"kubernetes.io/projected/a7eb77ab-add0-4601-a279-3f8fc956b52f-kube-api-access-v2zqr\") pod \"oauth-openshift-558db77b4-l5ld8\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.414203 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.424420 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9d9v8\" (UniqueName: \"kubernetes.io/projected/76dddf3f-2b2c-4644-ab68-402fa1e6b4d3-kube-api-access-9d9v8\") pod \"openshift-apiserver-operator-796bbdcf4f-zv6dz\" (UID: \"76dddf3f-2b2c-4644-ab68-402fa1e6b4d3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zv6dz" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.435335 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w9xq\" (UniqueName: \"kubernetes.io/projected/66cd34f5-c1e9-4b01-80ba-71b574b97ced-kube-api-access-2w9xq\") pod \"cluster-samples-operator-665b6dd947-qr8wb\" (UID: \"66cd34f5-c1e9-4b01-80ba-71b574b97ced\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qr8wb" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.457172 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.465833 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh8zh\" (UniqueName: \"kubernetes.io/projected/b18b84da-1277-4a24-b434-c93eb83a3af8-kube-api-access-jh8zh\") pod \"controller-manager-879f6c89f-tqhnj\" (UID: \"b18b84da-1277-4a24-b434-c93eb83a3af8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.469170 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.495583 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smn5q\" (UniqueName: \"kubernetes.io/projected/c3400b7c-5666-4b36-a50f-7800f6737527-kube-api-access-smn5q\") pod \"authentication-operator-69f744f599-bxplc\" (UID: \"c3400b7c-5666-4b36-a50f-7800f6737527\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.497202 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.517517 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.537735 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.557946 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.576839 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.596428 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.614102 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.618587 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.627267 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-wgtx2"] Dec 09 09:49:41 crc kubenswrapper[4824]: W1209 09:49:41.633302 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd6c14a0_99d8_485d_81f3_cc4ebe5a943b.slice/crio-360f84057d32ba228af5aacae31bc6cca680a0c254cdc94579eaa781190e0ab2 WatchSource:0}: Error finding container 360f84057d32ba228af5aacae31bc6cca680a0c254cdc94579eaa781190e0ab2: Status 404 returned error can't find the container with id 360f84057d32ba228af5aacae31bc6cca680a0c254cdc94579eaa781190e0ab2 Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.634877 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.652366 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dpmx\" (UniqueName: \"kubernetes.io/projected/00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53-kube-api-access-4dpmx\") pod \"ingress-operator-5b745b69d9-jgwrj\" (UID: \"00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jgwrj" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.668612 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qr8wb" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.668829 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd"] Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.670326 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6042da41-13ef-460e-aee4-491c7758f635-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2t6n7\" (UID: \"6042da41-13ef-460e-aee4-491c7758f635\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2t6n7" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.695282 4824 request.go:700] Waited for 1.913284177s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console/serviceaccounts/console/token Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.695510 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zv6dz" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.698589 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53-bound-sa-token\") pod \"ingress-operator-5b745b69d9-jgwrj\" (UID: \"00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jgwrj" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.703687 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.719484 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdcvb\" (UniqueName: \"kubernetes.io/projected/0c5e7133-21dc-46e0-b334-9d36c0427aa3-kube-api-access-zdcvb\") pod \"console-f9d7485db-ldw7r\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.735428 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtwxj\" (UniqueName: \"kubernetes.io/projected/76df4be1-7bf3-461d-a9d7-6a2be7c0e425-kube-api-access-vtwxj\") pod \"machine-config-operator-74547568cd-k4fpg\" (UID: \"76df4be1-7bf3-461d-a9d7-6a2be7c0e425\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k4fpg" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.760762 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t985r\" (UniqueName: \"kubernetes.io/projected/5047f85e-0349-40fc-b966-1956bd66cbbc-kube-api-access-t985r\") pod \"kube-storage-version-migrator-operator-b67b599dd-x7j5h\" (UID: \"5047f85e-0349-40fc-b966-1956bd66cbbc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x7j5h" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.765695 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.776446 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6km2\" (UniqueName: \"kubernetes.io/projected/f4616db0-42ba-4a3b-88f2-57a0663d0a15-kube-api-access-k6km2\") pod \"etcd-operator-b45778765-7b9dd\" (UID: \"f4616db0-42ba-4a3b-88f2-57a0663d0a15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b9dd" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.785060 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.802025 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.813562 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv"] Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.816742 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.838115 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.859646 4824 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.866977 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-7b9dd" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.879118 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.904232 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:41 crc kubenswrapper[4824]: I1209 09:49:41.980604 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zv6dz"] Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.007935 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tqhnj"] Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.257962 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jgwrj" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.258771 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k4fpg" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.259111 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2t6n7" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.259357 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x7j5h" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.264365 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.264515 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6d910d95-4961-4d28-9599-087153ce7d2f-registry-tls\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.264630 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6d910d95-4961-4d28-9599-087153ce7d2f-registry-certificates\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.264720 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6d910d95-4961-4d28-9599-087153ce7d2f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.266057 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-bxplc"] Dec 09 09:49:42 crc kubenswrapper[4824]: E1209 09:49:42.267300 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:42.76727 +0000 UTC m=+139.101774667 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.272462 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qr8wb"] Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.275995 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-l5ld8"] Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.365906 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.366138 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/a4e4cf9c-48df-43b1-8ad7-9826c6220a45-machine-approver-tls\") pod \"machine-approver-56656f9798-mjsbv\" (UID: \"a4e4cf9c-48df-43b1-8ad7-9826c6220a45\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjsbv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.366218 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6d910d95-4961-4d28-9599-087153ce7d2f-registry-tls\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.366249 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncvbc\" (UniqueName: \"kubernetes.io/projected/7b26f121-04f6-4501-84b2-1833b927aa14-kube-api-access-ncvbc\") pod \"openshift-config-operator-7777fb866f-j4zxt\" (UID: \"7b26f121-04f6-4501-84b2-1833b927aa14\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.366271 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/7bff53ba-5406-445c-afcb-3247516a1258-audit\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.366309 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bff53ba-5406-445c-afcb-3247516a1258-config\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.366352 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14ebfe85-ea7c-49a4-a2e2-f8822c3a111e-config\") pod \"machine-api-operator-5694c8668f-4rv7m\" (UID: \"14ebfe85-ea7c-49a4-a2e2-f8822c3a111e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4rv7m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.366395 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/14ebfe85-ea7c-49a4-a2e2-f8822c3a111e-images\") pod \"machine-api-operator-5694c8668f-4rv7m\" (UID: \"14ebfe85-ea7c-49a4-a2e2-f8822c3a111e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4rv7m" Dec 09 09:49:42 crc kubenswrapper[4824]: E1209 09:49:42.367674 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:42.867636874 +0000 UTC m=+139.202141531 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.367727 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6d910d95-4961-4d28-9599-087153ce7d2f-bound-sa-token\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.367758 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7bff53ba-5406-445c-afcb-3247516a1258-encryption-config\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.367775 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b26f121-04f6-4501-84b2-1833b927aa14-serving-cert\") pod \"openshift-config-operator-7777fb866f-j4zxt\" (UID: \"7b26f121-04f6-4501-84b2-1833b927aa14\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.367839 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4e4cf9c-48df-43b1-8ad7-9826c6220a45-config\") pod \"machine-approver-56656f9798-mjsbv\" (UID: \"a4e4cf9c-48df-43b1-8ad7-9826c6220a45\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjsbv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.367911 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bff53ba-5406-445c-afcb-3247516a1258-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.367941 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92w67\" (UniqueName: \"kubernetes.io/projected/9966a4dc-9922-434f-8f3e-6efc5db8445e-kube-api-access-92w67\") pod \"dns-operator-744455d44c-c6559\" (UID: \"9966a4dc-9922-434f-8f3e-6efc5db8445e\") " pod="openshift-dns-operator/dns-operator-744455d44c-c6559" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.367975 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/06fc7061-cbbe-459e-bcde-09d1c700ceba-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-847gg\" (UID: \"06fc7061-cbbe-459e-bcde-09d1c700ceba\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-847gg" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.368000 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6d910d95-4961-4d28-9599-087153ce7d2f-registry-certificates\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.368020 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpkmk\" (UniqueName: \"kubernetes.io/projected/7bff53ba-5406-445c-afcb-3247516a1258-kube-api-access-fpkmk\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.368038 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d6vn\" (UniqueName: \"kubernetes.io/projected/a4e4cf9c-48df-43b1-8ad7-9826c6220a45-kube-api-access-5d6vn\") pod \"machine-approver-56656f9798-mjsbv\" (UID: \"a4e4cf9c-48df-43b1-8ad7-9826c6220a45\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjsbv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.368064 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7bff53ba-5406-445c-afcb-3247516a1258-etcd-serving-ca\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.368081 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9966a4dc-9922-434f-8f3e-6efc5db8445e-metrics-tls\") pod \"dns-operator-744455d44c-c6559\" (UID: \"9966a4dc-9922-434f-8f3e-6efc5db8445e\") " pod="openshift-dns-operator/dns-operator-744455d44c-c6559" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.368149 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6d910d95-4961-4d28-9599-087153ce7d2f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.368167 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/06fc7061-cbbe-459e-bcde-09d1c700ceba-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-847gg\" (UID: \"06fc7061-cbbe-459e-bcde-09d1c700ceba\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-847gg" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.368223 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/7bff53ba-5406-445c-afcb-3247516a1258-image-import-ca\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.368273 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/14ebfe85-ea7c-49a4-a2e2-f8822c3a111e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4rv7m\" (UID: \"14ebfe85-ea7c-49a4-a2e2-f8822c3a111e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4rv7m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.368297 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bfj4\" (UniqueName: \"kubernetes.io/projected/14ebfe85-ea7c-49a4-a2e2-f8822c3a111e-kube-api-access-5bfj4\") pod \"machine-api-operator-5694c8668f-4rv7m\" (UID: \"14ebfe85-ea7c-49a4-a2e2-f8822c3a111e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4rv7m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.369352 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6d910d95-4961-4d28-9599-087153ce7d2f-registry-certificates\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.369423 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/06fc7061-cbbe-459e-bcde-09d1c700ceba-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-847gg\" (UID: \"06fc7061-cbbe-459e-bcde-09d1c700ceba\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-847gg" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.369462 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6d910d95-4961-4d28-9599-087153ce7d2f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.369492 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7bff53ba-5406-445c-afcb-3247516a1258-audit-dir\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.369509 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfrn2\" (UniqueName: \"kubernetes.io/projected/06fc7061-cbbe-459e-bcde-09d1c700ceba-kube-api-access-dfrn2\") pod \"cluster-image-registry-operator-dc59b4c8b-847gg\" (UID: \"06fc7061-cbbe-459e-bcde-09d1c700ceba\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-847gg" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.369533 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a4e4cf9c-48df-43b1-8ad7-9826c6220a45-auth-proxy-config\") pod \"machine-approver-56656f9798-mjsbv\" (UID: \"a4e4cf9c-48df-43b1-8ad7-9826c6220a45\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjsbv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.369554 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjzn9\" (UniqueName: \"kubernetes.io/projected/6efb8245-07e9-4c31-b8ab-53ab3685e593-kube-api-access-mjzn9\") pod \"downloads-7954f5f757-b9bnh\" (UID: \"6efb8245-07e9-4c31-b8ab-53ab3685e593\") " pod="openshift-console/downloads-7954f5f757-b9bnh" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.369752 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7bff53ba-5406-445c-afcb-3247516a1258-serving-cert\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.369890 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6d910d95-4961-4d28-9599-087153ce7d2f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.370208 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6d910d95-4961-4d28-9599-087153ce7d2f-trusted-ca\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.370398 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kthx2\" (UniqueName: \"kubernetes.io/projected/6d910d95-4961-4d28-9599-087153ce7d2f-kube-api-access-kthx2\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.370495 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.370576 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7bff53ba-5406-445c-afcb-3247516a1258-etcd-client\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.370807 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7bff53ba-5406-445c-afcb-3247516a1258-node-pullsecrets\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: E1209 09:49:42.370856 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:42.870839139 +0000 UTC m=+139.205343876 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.370893 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7b26f121-04f6-4501-84b2-1833b927aa14-available-featuregates\") pod \"openshift-config-operator-7777fb866f-j4zxt\" (UID: \"7b26f121-04f6-4501-84b2-1833b927aa14\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.377661 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6d910d95-4961-4d28-9599-087153ce7d2f-registry-tls\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.471605 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:42 crc kubenswrapper[4824]: E1209 09:49:42.471749 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:42.97172246 +0000 UTC m=+139.306227127 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.472290 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/648e0288-ccba-422a-a8c8-b41d12a20845-serving-cert\") pod \"service-ca-operator-777779d784-pk58m\" (UID: \"648e0288-ccba-422a-a8c8-b41d12a20845\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pk58m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.472322 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/45732d49-ce14-439b-8b7a-fb9fdf267fb1-webhook-cert\") pod \"packageserver-d55dfcdfc-tg5lv\" (UID: \"45732d49-ce14-439b-8b7a-fb9fdf267fb1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.472352 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/43d3560f-ad7b-4037-a980-0ed995e7e1d2-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9jtv8\" (UID: \"43d3560f-ad7b-4037-a980-0ed995e7e1d2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9jtv8" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.472375 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/dac1ac9b-4ddd-41b2-813b-ed7d8a71642f-default-certificate\") pod \"router-default-5444994796-mj7jx\" (UID: \"dac1ac9b-4ddd-41b2-813b-ed7d8a71642f\") " pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.472400 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb2d4bd5-8dac-4b9d-9125-51fca045c2b3-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-q8c9m\" (UID: \"eb2d4bd5-8dac-4b9d-9125-51fca045c2b3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q8c9m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.472429 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kthx2\" (UniqueName: \"kubernetes.io/projected/6d910d95-4961-4d28-9599-087153ce7d2f-kube-api-access-kthx2\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.472453 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/401617ee-8306-4176-9af1-4ed2a5f9af0d-cert\") pod \"ingress-canary-c9nzk\" (UID: \"401617ee-8306-4176-9af1-4ed2a5f9af0d\") " pod="openshift-ingress-canary/ingress-canary-c9nzk" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.472480 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2wk8\" (UniqueName: \"kubernetes.io/projected/54818ac0-7b5e-452f-b2f7-6a87a667a76f-kube-api-access-n2wk8\") pod \"marketplace-operator-79b997595-kv8sn\" (UID: \"54818ac0-7b5e-452f-b2f7-6a87a667a76f\") " pod="openshift-marketplace/marketplace-operator-79b997595-kv8sn" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.472507 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfjh9\" (UniqueName: \"kubernetes.io/projected/2474e03a-283f-48af-ac3a-eebe08168b8f-kube-api-access-xfjh9\") pod \"dns-default-zjjp6\" (UID: \"2474e03a-283f-48af-ac3a-eebe08168b8f\") " pod="openshift-dns/dns-default-zjjp6" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.472537 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7bff53ba-5406-445c-afcb-3247516a1258-etcd-client\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.472561 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/648e0288-ccba-422a-a8c8-b41d12a20845-config\") pod \"service-ca-operator-777779d784-pk58m\" (UID: \"648e0288-ccba-422a-a8c8-b41d12a20845\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pk58m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.473298 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/6c10edf9-cf75-4633-8d4c-9bbb9d9b6339-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-k8bgl\" (UID: \"6c10edf9-cf75-4633-8d4c-9bbb9d9b6339\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.473349 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/564831de-2ff3-4051-a1a0-c5317bf10ca1-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gmk4v\" (UID: \"564831de-2ff3-4051-a1a0-c5317bf10ca1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmk4v" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.473376 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/a4e4cf9c-48df-43b1-8ad7-9826c6220a45-machine-approver-tls\") pod \"machine-approver-56656f9798-mjsbv\" (UID: \"a4e4cf9c-48df-43b1-8ad7-9826c6220a45\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjsbv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.473405 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab5c6701-8122-4d9e-829d-93018eb67d7b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-pvgjx\" (UID: \"ab5c6701-8122-4d9e-829d-93018eb67d7b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pvgjx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.473432 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a7c02c2b-70ae-4b6f-85ae-5a38b737a73b-srv-cert\") pod \"olm-operator-6b444d44fb-c2dfm\" (UID: \"a7c02c2b-70ae-4b6f-85ae-5a38b737a73b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c2dfm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.473614 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbsxw\" (UniqueName: \"kubernetes.io/projected/43d3560f-ad7b-4037-a980-0ed995e7e1d2-kube-api-access-qbsxw\") pod \"control-plane-machine-set-operator-78cbb6b69f-9jtv8\" (UID: \"43d3560f-ad7b-4037-a980-0ed995e7e1d2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9jtv8" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.473652 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2474e03a-283f-48af-ac3a-eebe08168b8f-config-volume\") pod \"dns-default-zjjp6\" (UID: \"2474e03a-283f-48af-ac3a-eebe08168b8f\") " pod="openshift-dns/dns-default-zjjp6" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.473789 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/14ebfe85-ea7c-49a4-a2e2-f8822c3a111e-images\") pod \"machine-api-operator-5694c8668f-4rv7m\" (UID: \"14ebfe85-ea7c-49a4-a2e2-f8822c3a111e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4rv7m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.473828 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab5c6701-8122-4d9e-829d-93018eb67d7b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-pvgjx\" (UID: \"ab5c6701-8122-4d9e-829d-93018eb67d7b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pvgjx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.473877 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b26f121-04f6-4501-84b2-1833b927aa14-serving-cert\") pod \"openshift-config-operator-7777fb866f-j4zxt\" (UID: \"7b26f121-04f6-4501-84b2-1833b927aa14\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.473897 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5d1a49e0-07ed-4b75-8ce1-f6c8e9930312-node-bootstrap-token\") pod \"machine-config-server-g2cbw\" (UID: \"5d1a49e0-07ed-4b75-8ce1-f6c8e9930312\") " pod="openshift-machine-config-operator/machine-config-server-g2cbw" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.473925 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgqnn\" (UniqueName: \"kubernetes.io/projected/5d1a49e0-07ed-4b75-8ce1-f6c8e9930312-kube-api-access-mgqnn\") pod \"machine-config-server-g2cbw\" (UID: \"5d1a49e0-07ed-4b75-8ce1-f6c8e9930312\") " pod="openshift-machine-config-operator/machine-config-server-g2cbw" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.473944 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4e4cf9c-48df-43b1-8ad7-9826c6220a45-config\") pod \"machine-approver-56656f9798-mjsbv\" (UID: \"a4e4cf9c-48df-43b1-8ad7-9826c6220a45\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjsbv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.473966 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dac1ac9b-4ddd-41b2-813b-ed7d8a71642f-service-ca-bundle\") pod \"router-default-5444994796-mj7jx\" (UID: \"dac1ac9b-4ddd-41b2-813b-ed7d8a71642f\") " pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.474001 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bff53ba-5406-445c-afcb-3247516a1258-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.474022 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/54818ac0-7b5e-452f-b2f7-6a87a667a76f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kv8sn\" (UID: \"54818ac0-7b5e-452f-b2f7-6a87a667a76f\") " pod="openshift-marketplace/marketplace-operator-79b997595-kv8sn" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.474047 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/06fc7061-cbbe-459e-bcde-09d1c700ceba-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-847gg\" (UID: \"06fc7061-cbbe-459e-bcde-09d1c700ceba\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-847gg" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.474072 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92w67\" (UniqueName: \"kubernetes.io/projected/9966a4dc-9922-434f-8f3e-6efc5db8445e-kube-api-access-92w67\") pod \"dns-operator-744455d44c-c6559\" (UID: \"9966a4dc-9922-434f-8f3e-6efc5db8445e\") " pod="openshift-dns-operator/dns-operator-744455d44c-c6559" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.474095 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjlmb\" (UniqueName: \"kubernetes.io/projected/ab5c6701-8122-4d9e-829d-93018eb67d7b-kube-api-access-hjlmb\") pod \"openshift-controller-manager-operator-756b6f6bc6-pvgjx\" (UID: \"ab5c6701-8122-4d9e-829d-93018eb67d7b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pvgjx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.474121 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpkmk\" (UniqueName: \"kubernetes.io/projected/7bff53ba-5406-445c-afcb-3247516a1258-kube-api-access-fpkmk\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.474148 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb2d4bd5-8dac-4b9d-9125-51fca045c2b3-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-q8c9m\" (UID: \"eb2d4bd5-8dac-4b9d-9125-51fca045c2b3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q8c9m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.474183 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5d1a49e0-07ed-4b75-8ce1-f6c8e9930312-certs\") pod \"machine-config-server-g2cbw\" (UID: \"5d1a49e0-07ed-4b75-8ce1-f6c8e9930312\") " pod="openshift-machine-config-operator/machine-config-server-g2cbw" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.474235 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/055ead87-3e57-4053-a026-24566879437c-config-volume\") pod \"collect-profiles-29421225-96fgb\" (UID: \"055ead87-3e57-4053-a026-24566879437c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421225-96fgb" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.474743 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/14ebfe85-ea7c-49a4-a2e2-f8822c3a111e-images\") pod \"machine-api-operator-5694c8668f-4rv7m\" (UID: \"14ebfe85-ea7c-49a4-a2e2-f8822c3a111e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4rv7m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.474967 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4e4cf9c-48df-43b1-8ad7-9826c6220a45-config\") pod \"machine-approver-56656f9798-mjsbv\" (UID: \"a4e4cf9c-48df-43b1-8ad7-9826c6220a45\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjsbv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.476412 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6d910d95-4961-4d28-9599-087153ce7d2f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.476450 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/055ead87-3e57-4053-a026-24566879437c-secret-volume\") pod \"collect-profiles-29421225-96fgb\" (UID: \"055ead87-3e57-4053-a026-24566879437c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421225-96fgb" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.476477 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dac1ac9b-4ddd-41b2-813b-ed7d8a71642f-metrics-certs\") pod \"router-default-5444994796-mj7jx\" (UID: \"dac1ac9b-4ddd-41b2-813b-ed7d8a71642f\") " pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.477703 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bff53ba-5406-445c-afcb-3247516a1258-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.478563 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfvr4\" (UniqueName: \"kubernetes.io/projected/0f899ebd-d4e0-4e7c-82ea-914abad9dab6-kube-api-access-tfvr4\") pod \"migrator-59844c95c7-zl4mr\" (UID: \"0f899ebd-d4e0-4e7c-82ea-914abad9dab6\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zl4mr" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.478602 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/67324e78-0fdb-4c67-a755-e2277801c106-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-mnksj\" (UID: \"67324e78-0fdb-4c67-a755-e2277801c106\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mnksj" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.478652 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/dac1ac9b-4ddd-41b2-813b-ed7d8a71642f-stats-auth\") pod \"router-default-5444994796-mj7jx\" (UID: \"dac1ac9b-4ddd-41b2-813b-ed7d8a71642f\") " pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.478726 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-td5l8\" (UniqueName: \"kubernetes.io/projected/7a94cb6d-599c-4dae-b7db-43f67a6620b6-kube-api-access-td5l8\") pod \"service-ca-9c57cc56f-99q6g\" (UID: \"7a94cb6d-599c-4dae-b7db-43f67a6620b6\") " pod="openshift-service-ca/service-ca-9c57cc56f-99q6g" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.478745 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/45732d49-ce14-439b-8b7a-fb9fdf267fb1-apiservice-cert\") pod \"packageserver-d55dfcdfc-tg5lv\" (UID: \"45732d49-ce14-439b-8b7a-fb9fdf267fb1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.478810 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7bff53ba-5406-445c-afcb-3247516a1258-audit-dir\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.478836 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a4e4cf9c-48df-43b1-8ad7-9826c6220a45-auth-proxy-config\") pod \"machine-approver-56656f9798-mjsbv\" (UID: \"a4e4cf9c-48df-43b1-8ad7-9826c6220a45\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjsbv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.478891 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjzn9\" (UniqueName: \"kubernetes.io/projected/6efb8245-07e9-4c31-b8ab-53ab3685e593-kube-api-access-mjzn9\") pod \"downloads-7954f5f757-b9bnh\" (UID: \"6efb8245-07e9-4c31-b8ab-53ab3685e593\") " pod="openshift-console/downloads-7954f5f757-b9bnh" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.478913 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7bff53ba-5406-445c-afcb-3247516a1258-serving-cert\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.478939 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2474e03a-283f-48af-ac3a-eebe08168b8f-metrics-tls\") pod \"dns-default-zjjp6\" (UID: \"2474e03a-283f-48af-ac3a-eebe08168b8f\") " pod="openshift-dns/dns-default-zjjp6" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.479468 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a4e4cf9c-48df-43b1-8ad7-9826c6220a45-auth-proxy-config\") pod \"machine-approver-56656f9798-mjsbv\" (UID: \"a4e4cf9c-48df-43b1-8ad7-9826c6220a45\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjsbv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.479751 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/a4e4cf9c-48df-43b1-8ad7-9826c6220a45-machine-approver-tls\") pod \"machine-approver-56656f9798-mjsbv\" (UID: \"a4e4cf9c-48df-43b1-8ad7-9826c6220a45\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjsbv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.480496 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6d910d95-4961-4d28-9599-087153ce7d2f-trusted-ca\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.480663 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.480809 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g2fd\" (UniqueName: \"kubernetes.io/projected/37c820ff-b4bf-41ae-bcf3-8e0f7368bb12-kube-api-access-6g2fd\") pod \"machine-config-controller-84d6567774-hjg74\" (UID: \"37c820ff-b4bf-41ae-bcf3-8e0f7368bb12\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hjg74" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.480844 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7a94cb6d-599c-4dae-b7db-43f67a6620b6-signing-cabundle\") pod \"service-ca-9c57cc56f-99q6g\" (UID: \"7a94cb6d-599c-4dae-b7db-43f67a6620b6\") " pod="openshift-service-ca/service-ca-9c57cc56f-99q6g" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.480883 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/54818ac0-7b5e-452f-b2f7-6a87a667a76f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kv8sn\" (UID: \"54818ac0-7b5e-452f-b2f7-6a87a667a76f\") " pod="openshift-marketplace/marketplace-operator-79b997595-kv8sn" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.480913 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a7c02c2b-70ae-4b6f-85ae-5a38b737a73b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-c2dfm\" (UID: \"a7c02c2b-70ae-4b6f-85ae-5a38b737a73b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c2dfm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.481022 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7bff53ba-5406-445c-afcb-3247516a1258-node-pullsecrets\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.481049 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7b26f121-04f6-4501-84b2-1833b927aa14-available-featuregates\") pod \"openshift-config-operator-7777fb866f-j4zxt\" (UID: \"7b26f121-04f6-4501-84b2-1833b927aa14\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.481072 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/62823972-71dc-4f7d-b4f6-da0cb90dcf36-profile-collector-cert\") pod \"catalog-operator-68c6474976-9c2mr\" (UID: \"62823972-71dc-4f7d-b4f6-da0cb90dcf36\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.481266 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7bff53ba-5406-445c-afcb-3247516a1258-audit-dir\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.481303 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6d910d95-4961-4d28-9599-087153ce7d2f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.481333 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7bff53ba-5406-445c-afcb-3247516a1258-node-pullsecrets\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.481388 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/37c820ff-b4bf-41ae-bcf3-8e0f7368bb12-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-hjg74\" (UID: \"37c820ff-b4bf-41ae-bcf3-8e0f7368bb12\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hjg74" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.481644 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7b26f121-04f6-4501-84b2-1833b927aa14-available-featuregates\") pod \"openshift-config-operator-7777fb866f-j4zxt\" (UID: \"7b26f121-04f6-4501-84b2-1833b927aa14\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.481698 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhjpz\" (UniqueName: \"kubernetes.io/projected/a7c02c2b-70ae-4b6f-85ae-5a38b737a73b-kube-api-access-vhjpz\") pod \"olm-operator-6b444d44fb-c2dfm\" (UID: \"a7c02c2b-70ae-4b6f-85ae-5a38b737a73b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c2dfm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.481744 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncvbc\" (UniqueName: \"kubernetes.io/projected/7b26f121-04f6-4501-84b2-1833b927aa14-kube-api-access-ncvbc\") pod \"openshift-config-operator-7777fb866f-j4zxt\" (UID: \"7b26f121-04f6-4501-84b2-1833b927aa14\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.481770 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/7bff53ba-5406-445c-afcb-3247516a1258-audit\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.481805 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/564831de-2ff3-4051-a1a0-c5317bf10ca1-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gmk4v\" (UID: \"564831de-2ff3-4051-a1a0-c5317bf10ca1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmk4v" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.481847 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bff53ba-5406-445c-afcb-3247516a1258-config\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.481865 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14ebfe85-ea7c-49a4-a2e2-f8822c3a111e-config\") pod \"machine-api-operator-5694c8668f-4rv7m\" (UID: \"14ebfe85-ea7c-49a4-a2e2-f8822c3a111e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4rv7m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.481884 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/37c820ff-b4bf-41ae-bcf3-8e0f7368bb12-proxy-tls\") pod \"machine-config-controller-84d6567774-hjg74\" (UID: \"37c820ff-b4bf-41ae-bcf3-8e0f7368bb12\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hjg74" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.481903 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxqf8\" (UniqueName: \"kubernetes.io/projected/dac1ac9b-4ddd-41b2-813b-ed7d8a71642f-kube-api-access-fxqf8\") pod \"router-default-5444994796-mj7jx\" (UID: \"dac1ac9b-4ddd-41b2-813b-ed7d8a71642f\") " pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.484079 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/7bff53ba-5406-445c-afcb-3247516a1258-audit\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.484502 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bff53ba-5406-445c-afcb-3247516a1258-config\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.485084 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14ebfe85-ea7c-49a4-a2e2-f8822c3a111e-config\") pod \"machine-api-operator-5694c8668f-4rv7m\" (UID: \"14ebfe85-ea7c-49a4-a2e2-f8822c3a111e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4rv7m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.485144 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6d910d95-4961-4d28-9599-087153ce7d2f-bound-sa-token\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.485189 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7bff53ba-5406-445c-afcb-3247516a1258-encryption-config\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.485214 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hglrf\" (UniqueName: \"kubernetes.io/projected/648e0288-ccba-422a-a8c8-b41d12a20845-kube-api-access-hglrf\") pod \"service-ca-operator-777779d784-pk58m\" (UID: \"648e0288-ccba-422a-a8c8-b41d12a20845\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pk58m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.485260 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lv67\" (UniqueName: \"kubernetes.io/projected/401617ee-8306-4176-9af1-4ed2a5f9af0d-kube-api-access-8lv67\") pod \"ingress-canary-c9nzk\" (UID: \"401617ee-8306-4176-9af1-4ed2a5f9af0d\") " pod="openshift-ingress-canary/ingress-canary-c9nzk" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.485278 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/564831de-2ff3-4051-a1a0-c5317bf10ca1-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gmk4v\" (UID: \"564831de-2ff3-4051-a1a0-c5317bf10ca1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmk4v" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.486119 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg7lm\" (UniqueName: \"kubernetes.io/projected/c0ac631f-5d13-4814-84d3-61f7bd704f9e-kube-api-access-qg7lm\") pod \"csi-hostpathplugin-vsdpm\" (UID: \"c0ac631f-5d13-4814-84d3-61f7bd704f9e\") " pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.486205 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c0ac631f-5d13-4814-84d3-61f7bd704f9e-registration-dir\") pod \"csi-hostpathplugin-vsdpm\" (UID: \"c0ac631f-5d13-4814-84d3-61f7bd704f9e\") " pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.486228 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/45732d49-ce14-439b-8b7a-fb9fdf267fb1-tmpfs\") pod \"packageserver-d55dfcdfc-tg5lv\" (UID: \"45732d49-ce14-439b-8b7a-fb9fdf267fb1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.486310 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c0ac631f-5d13-4814-84d3-61f7bd704f9e-mountpoint-dir\") pod \"csi-hostpathplugin-vsdpm\" (UID: \"c0ac631f-5d13-4814-84d3-61f7bd704f9e\") " pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.486356 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7bff53ba-5406-445c-afcb-3247516a1258-etcd-serving-ca\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.486379 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d6vn\" (UniqueName: \"kubernetes.io/projected/a4e4cf9c-48df-43b1-8ad7-9826c6220a45-kube-api-access-5d6vn\") pod \"machine-approver-56656f9798-mjsbv\" (UID: \"a4e4cf9c-48df-43b1-8ad7-9826c6220a45\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjsbv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.486399 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9966a4dc-9922-434f-8f3e-6efc5db8445e-metrics-tls\") pod \"dns-operator-744455d44c-c6559\" (UID: \"9966a4dc-9922-434f-8f3e-6efc5db8445e\") " pod="openshift-dns-operator/dns-operator-744455d44c-c6559" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.486420 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb2d4bd5-8dac-4b9d-9125-51fca045c2b3-config\") pod \"kube-controller-manager-operator-78b949d7b-q8c9m\" (UID: \"eb2d4bd5-8dac-4b9d-9125-51fca045c2b3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q8c9m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.486510 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/62823972-71dc-4f7d-b4f6-da0cb90dcf36-srv-cert\") pod \"catalog-operator-68c6474976-9c2mr\" (UID: \"62823972-71dc-4f7d-b4f6-da0cb90dcf36\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.486533 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkw4k\" (UniqueName: \"kubernetes.io/projected/45732d49-ce14-439b-8b7a-fb9fdf267fb1-kube-api-access-vkw4k\") pod \"packageserver-d55dfcdfc-tg5lv\" (UID: \"45732d49-ce14-439b-8b7a-fb9fdf267fb1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.486615 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/06fc7061-cbbe-459e-bcde-09d1c700ceba-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-847gg\" (UID: \"06fc7061-cbbe-459e-bcde-09d1c700ceba\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-847gg" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.486641 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7a94cb6d-599c-4dae-b7db-43f67a6620b6-signing-key\") pod \"service-ca-9c57cc56f-99q6g\" (UID: \"7a94cb6d-599c-4dae-b7db-43f67a6620b6\") " pod="openshift-service-ca/service-ca-9c57cc56f-99q6g" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.486662 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c0ac631f-5d13-4814-84d3-61f7bd704f9e-plugins-dir\") pod \"csi-hostpathplugin-vsdpm\" (UID: \"c0ac631f-5d13-4814-84d3-61f7bd704f9e\") " pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.488337 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7bff53ba-5406-445c-afcb-3247516a1258-etcd-serving-ca\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.488547 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b26f121-04f6-4501-84b2-1833b927aa14-serving-cert\") pod \"openshift-config-operator-7777fb866f-j4zxt\" (UID: \"7b26f121-04f6-4501-84b2-1833b927aa14\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.488752 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/06fc7061-cbbe-459e-bcde-09d1c700ceba-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-847gg\" (UID: \"06fc7061-cbbe-459e-bcde-09d1c700ceba\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-847gg" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.488939 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/7bff53ba-5406-445c-afcb-3247516a1258-image-import-ca\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.488963 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7jwh\" (UniqueName: \"kubernetes.io/projected/62823972-71dc-4f7d-b4f6-da0cb90dcf36-kube-api-access-k7jwh\") pod \"catalog-operator-68c6474976-9c2mr\" (UID: \"62823972-71dc-4f7d-b4f6-da0cb90dcf36\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.489031 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qwml\" (UniqueName: \"kubernetes.io/projected/055ead87-3e57-4053-a026-24566879437c-kube-api-access-2qwml\") pod \"collect-profiles-29421225-96fgb\" (UID: \"055ead87-3e57-4053-a026-24566879437c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421225-96fgb" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.489061 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlgpp\" (UniqueName: \"kubernetes.io/projected/6c10edf9-cf75-4633-8d4c-9bbb9d9b6339-kube-api-access-mlgpp\") pod \"package-server-manager-789f6589d5-k8bgl\" (UID: \"6c10edf9-cf75-4633-8d4c-9bbb9d9b6339\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.489087 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/14ebfe85-ea7c-49a4-a2e2-f8822c3a111e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4rv7m\" (UID: \"14ebfe85-ea7c-49a4-a2e2-f8822c3a111e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4rv7m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.489106 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bfj4\" (UniqueName: \"kubernetes.io/projected/14ebfe85-ea7c-49a4-a2e2-f8822c3a111e-kube-api-access-5bfj4\") pod \"machine-api-operator-5694c8668f-4rv7m\" (UID: \"14ebfe85-ea7c-49a4-a2e2-f8822c3a111e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4rv7m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.489172 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c0ac631f-5d13-4814-84d3-61f7bd704f9e-csi-data-dir\") pod \"csi-hostpathplugin-vsdpm\" (UID: \"c0ac631f-5d13-4814-84d3-61f7bd704f9e\") " pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.489191 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8q5s\" (UniqueName: \"kubernetes.io/projected/67324e78-0fdb-4c67-a755-e2277801c106-kube-api-access-k8q5s\") pod \"multus-admission-controller-857f4d67dd-mnksj\" (UID: \"67324e78-0fdb-4c67-a755-e2277801c106\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mnksj" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.489281 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/06fc7061-cbbe-459e-bcde-09d1c700ceba-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-847gg\" (UID: \"06fc7061-cbbe-459e-bcde-09d1c700ceba\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-847gg" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.489301 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfrn2\" (UniqueName: \"kubernetes.io/projected/06fc7061-cbbe-459e-bcde-09d1c700ceba-kube-api-access-dfrn2\") pod \"cluster-image-registry-operator-dc59b4c8b-847gg\" (UID: \"06fc7061-cbbe-459e-bcde-09d1c700ceba\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-847gg" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.489765 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6d910d95-4961-4d28-9599-087153ce7d2f-trusted-ca\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.491136 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7bff53ba-5406-445c-afcb-3247516a1258-encryption-config\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.492759 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/06fc7061-cbbe-459e-bcde-09d1c700ceba-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-847gg\" (UID: \"06fc7061-cbbe-459e-bcde-09d1c700ceba\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-847gg" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.492848 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c0ac631f-5d13-4814-84d3-61f7bd704f9e-socket-dir\") pod \"csi-hostpathplugin-vsdpm\" (UID: \"c0ac631f-5d13-4814-84d3-61f7bd704f9e\") " pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" Dec 09 09:49:42 crc kubenswrapper[4824]: E1209 09:49:42.492998 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:42.992977336 +0000 UTC m=+139.327482213 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.493528 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7bff53ba-5406-445c-afcb-3247516a1258-serving-cert\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.500465 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/7bff53ba-5406-445c-afcb-3247516a1258-image-import-ca\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.500792 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7bff53ba-5406-445c-afcb-3247516a1258-etcd-client\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.510209 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9966a4dc-9922-434f-8f3e-6efc5db8445e-metrics-tls\") pod \"dns-operator-744455d44c-c6559\" (UID: \"9966a4dc-9922-434f-8f3e-6efc5db8445e\") " pod="openshift-dns-operator/dns-operator-744455d44c-c6559" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.510710 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/14ebfe85-ea7c-49a4-a2e2-f8822c3a111e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4rv7m\" (UID: \"14ebfe85-ea7c-49a4-a2e2-f8822c3a111e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4rv7m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.519547 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kthx2\" (UniqueName: \"kubernetes.io/projected/6d910d95-4961-4d28-9599-087153ce7d2f-kube-api-access-kthx2\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.534988 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpkmk\" (UniqueName: \"kubernetes.io/projected/7bff53ba-5406-445c-afcb-3247516a1258-kube-api-access-fpkmk\") pod \"apiserver-76f77b778f-pd82d\" (UID: \"7bff53ba-5406-445c-afcb-3247516a1258\") " pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.557561 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92w67\" (UniqueName: \"kubernetes.io/projected/9966a4dc-9922-434f-8f3e-6efc5db8445e-kube-api-access-92w67\") pod \"dns-operator-744455d44c-c6559\" (UID: \"9966a4dc-9922-434f-8f3e-6efc5db8445e\") " pod="openshift-dns-operator/dns-operator-744455d44c-c6559" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.580037 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjzn9\" (UniqueName: \"kubernetes.io/projected/6efb8245-07e9-4c31-b8ab-53ab3685e593-kube-api-access-mjzn9\") pod \"downloads-7954f5f757-b9bnh\" (UID: \"6efb8245-07e9-4c31-b8ab-53ab3685e593\") " pod="openshift-console/downloads-7954f5f757-b9bnh" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.593431 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:42 crc kubenswrapper[4824]: E1209 09:49:42.593828 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:43.093769224 +0000 UTC m=+139.428273891 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.593902 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/648e0288-ccba-422a-a8c8-b41d12a20845-config\") pod \"service-ca-operator-777779d784-pk58m\" (UID: \"648e0288-ccba-422a-a8c8-b41d12a20845\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pk58m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.593942 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/6c10edf9-cf75-4633-8d4c-9bbb9d9b6339-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-k8bgl\" (UID: \"6c10edf9-cf75-4633-8d4c-9bbb9d9b6339\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.593961 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/564831de-2ff3-4051-a1a0-c5317bf10ca1-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gmk4v\" (UID: \"564831de-2ff3-4051-a1a0-c5317bf10ca1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmk4v" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.593980 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab5c6701-8122-4d9e-829d-93018eb67d7b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-pvgjx\" (UID: \"ab5c6701-8122-4d9e-829d-93018eb67d7b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pvgjx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.593998 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a7c02c2b-70ae-4b6f-85ae-5a38b737a73b-srv-cert\") pod \"olm-operator-6b444d44fb-c2dfm\" (UID: \"a7c02c2b-70ae-4b6f-85ae-5a38b737a73b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c2dfm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594023 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbsxw\" (UniqueName: \"kubernetes.io/projected/43d3560f-ad7b-4037-a980-0ed995e7e1d2-kube-api-access-qbsxw\") pod \"control-plane-machine-set-operator-78cbb6b69f-9jtv8\" (UID: \"43d3560f-ad7b-4037-a980-0ed995e7e1d2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9jtv8" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594044 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2474e03a-283f-48af-ac3a-eebe08168b8f-config-volume\") pod \"dns-default-zjjp6\" (UID: \"2474e03a-283f-48af-ac3a-eebe08168b8f\") " pod="openshift-dns/dns-default-zjjp6" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594068 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab5c6701-8122-4d9e-829d-93018eb67d7b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-pvgjx\" (UID: \"ab5c6701-8122-4d9e-829d-93018eb67d7b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pvgjx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594281 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5d1a49e0-07ed-4b75-8ce1-f6c8e9930312-node-bootstrap-token\") pod \"machine-config-server-g2cbw\" (UID: \"5d1a49e0-07ed-4b75-8ce1-f6c8e9930312\") " pod="openshift-machine-config-operator/machine-config-server-g2cbw" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594300 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgqnn\" (UniqueName: \"kubernetes.io/projected/5d1a49e0-07ed-4b75-8ce1-f6c8e9930312-kube-api-access-mgqnn\") pod \"machine-config-server-g2cbw\" (UID: \"5d1a49e0-07ed-4b75-8ce1-f6c8e9930312\") " pod="openshift-machine-config-operator/machine-config-server-g2cbw" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594319 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dac1ac9b-4ddd-41b2-813b-ed7d8a71642f-service-ca-bundle\") pod \"router-default-5444994796-mj7jx\" (UID: \"dac1ac9b-4ddd-41b2-813b-ed7d8a71642f\") " pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594353 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/54818ac0-7b5e-452f-b2f7-6a87a667a76f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kv8sn\" (UID: \"54818ac0-7b5e-452f-b2f7-6a87a667a76f\") " pod="openshift-marketplace/marketplace-operator-79b997595-kv8sn" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594390 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjlmb\" (UniqueName: \"kubernetes.io/projected/ab5c6701-8122-4d9e-829d-93018eb67d7b-kube-api-access-hjlmb\") pod \"openshift-controller-manager-operator-756b6f6bc6-pvgjx\" (UID: \"ab5c6701-8122-4d9e-829d-93018eb67d7b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pvgjx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594409 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb2d4bd5-8dac-4b9d-9125-51fca045c2b3-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-q8c9m\" (UID: \"eb2d4bd5-8dac-4b9d-9125-51fca045c2b3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q8c9m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594426 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5d1a49e0-07ed-4b75-8ce1-f6c8e9930312-certs\") pod \"machine-config-server-g2cbw\" (UID: \"5d1a49e0-07ed-4b75-8ce1-f6c8e9930312\") " pod="openshift-machine-config-operator/machine-config-server-g2cbw" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594444 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/055ead87-3e57-4053-a026-24566879437c-config-volume\") pod \"collect-profiles-29421225-96fgb\" (UID: \"055ead87-3e57-4053-a026-24566879437c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421225-96fgb" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594466 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/055ead87-3e57-4053-a026-24566879437c-secret-volume\") pod \"collect-profiles-29421225-96fgb\" (UID: \"055ead87-3e57-4053-a026-24566879437c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421225-96fgb" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594484 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dac1ac9b-4ddd-41b2-813b-ed7d8a71642f-metrics-certs\") pod \"router-default-5444994796-mj7jx\" (UID: \"dac1ac9b-4ddd-41b2-813b-ed7d8a71642f\") " pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594504 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfvr4\" (UniqueName: \"kubernetes.io/projected/0f899ebd-d4e0-4e7c-82ea-914abad9dab6-kube-api-access-tfvr4\") pod \"migrator-59844c95c7-zl4mr\" (UID: \"0f899ebd-d4e0-4e7c-82ea-914abad9dab6\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zl4mr" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594523 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/67324e78-0fdb-4c67-a755-e2277801c106-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-mnksj\" (UID: \"67324e78-0fdb-4c67-a755-e2277801c106\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mnksj" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594540 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/dac1ac9b-4ddd-41b2-813b-ed7d8a71642f-stats-auth\") pod \"router-default-5444994796-mj7jx\" (UID: \"dac1ac9b-4ddd-41b2-813b-ed7d8a71642f\") " pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594559 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-td5l8\" (UniqueName: \"kubernetes.io/projected/7a94cb6d-599c-4dae-b7db-43f67a6620b6-kube-api-access-td5l8\") pod \"service-ca-9c57cc56f-99q6g\" (UID: \"7a94cb6d-599c-4dae-b7db-43f67a6620b6\") " pod="openshift-service-ca/service-ca-9c57cc56f-99q6g" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594576 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/45732d49-ce14-439b-8b7a-fb9fdf267fb1-apiservice-cert\") pod \"packageserver-d55dfcdfc-tg5lv\" (UID: \"45732d49-ce14-439b-8b7a-fb9fdf267fb1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594605 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2474e03a-283f-48af-ac3a-eebe08168b8f-metrics-tls\") pod \"dns-default-zjjp6\" (UID: \"2474e03a-283f-48af-ac3a-eebe08168b8f\") " pod="openshift-dns/dns-default-zjjp6" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594632 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594668 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g2fd\" (UniqueName: \"kubernetes.io/projected/37c820ff-b4bf-41ae-bcf3-8e0f7368bb12-kube-api-access-6g2fd\") pod \"machine-config-controller-84d6567774-hjg74\" (UID: \"37c820ff-b4bf-41ae-bcf3-8e0f7368bb12\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hjg74" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594683 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7a94cb6d-599c-4dae-b7db-43f67a6620b6-signing-cabundle\") pod \"service-ca-9c57cc56f-99q6g\" (UID: \"7a94cb6d-599c-4dae-b7db-43f67a6620b6\") " pod="openshift-service-ca/service-ca-9c57cc56f-99q6g" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594694 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab5c6701-8122-4d9e-829d-93018eb67d7b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-pvgjx\" (UID: \"ab5c6701-8122-4d9e-829d-93018eb67d7b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pvgjx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594698 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/54818ac0-7b5e-452f-b2f7-6a87a667a76f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kv8sn\" (UID: \"54818ac0-7b5e-452f-b2f7-6a87a667a76f\") " pod="openshift-marketplace/marketplace-operator-79b997595-kv8sn" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594752 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a7c02c2b-70ae-4b6f-85ae-5a38b737a73b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-c2dfm\" (UID: \"a7c02c2b-70ae-4b6f-85ae-5a38b737a73b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c2dfm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594799 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/62823972-71dc-4f7d-b4f6-da0cb90dcf36-profile-collector-cert\") pod \"catalog-operator-68c6474976-9c2mr\" (UID: \"62823972-71dc-4f7d-b4f6-da0cb90dcf36\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594823 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/37c820ff-b4bf-41ae-bcf3-8e0f7368bb12-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-hjg74\" (UID: \"37c820ff-b4bf-41ae-bcf3-8e0f7368bb12\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hjg74" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594839 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhjpz\" (UniqueName: \"kubernetes.io/projected/a7c02c2b-70ae-4b6f-85ae-5a38b737a73b-kube-api-access-vhjpz\") pod \"olm-operator-6b444d44fb-c2dfm\" (UID: \"a7c02c2b-70ae-4b6f-85ae-5a38b737a73b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c2dfm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594861 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/564831de-2ff3-4051-a1a0-c5317bf10ca1-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gmk4v\" (UID: \"564831de-2ff3-4051-a1a0-c5317bf10ca1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmk4v" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594878 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/37c820ff-b4bf-41ae-bcf3-8e0f7368bb12-proxy-tls\") pod \"machine-config-controller-84d6567774-hjg74\" (UID: \"37c820ff-b4bf-41ae-bcf3-8e0f7368bb12\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hjg74" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594896 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxqf8\" (UniqueName: \"kubernetes.io/projected/dac1ac9b-4ddd-41b2-813b-ed7d8a71642f-kube-api-access-fxqf8\") pod \"router-default-5444994796-mj7jx\" (UID: \"dac1ac9b-4ddd-41b2-813b-ed7d8a71642f\") " pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594925 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hglrf\" (UniqueName: \"kubernetes.io/projected/648e0288-ccba-422a-a8c8-b41d12a20845-kube-api-access-hglrf\") pod \"service-ca-operator-777779d784-pk58m\" (UID: \"648e0288-ccba-422a-a8c8-b41d12a20845\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pk58m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594946 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lv67\" (UniqueName: \"kubernetes.io/projected/401617ee-8306-4176-9af1-4ed2a5f9af0d-kube-api-access-8lv67\") pod \"ingress-canary-c9nzk\" (UID: \"401617ee-8306-4176-9af1-4ed2a5f9af0d\") " pod="openshift-ingress-canary/ingress-canary-c9nzk" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.594960 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/564831de-2ff3-4051-a1a0-c5317bf10ca1-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gmk4v\" (UID: \"564831de-2ff3-4051-a1a0-c5317bf10ca1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmk4v" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596173 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/37c820ff-b4bf-41ae-bcf3-8e0f7368bb12-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-hjg74\" (UID: \"37c820ff-b4bf-41ae-bcf3-8e0f7368bb12\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hjg74" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596285 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg7lm\" (UniqueName: \"kubernetes.io/projected/c0ac631f-5d13-4814-84d3-61f7bd704f9e-kube-api-access-qg7lm\") pod \"csi-hostpathplugin-vsdpm\" (UID: \"c0ac631f-5d13-4814-84d3-61f7bd704f9e\") " pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596319 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c0ac631f-5d13-4814-84d3-61f7bd704f9e-registration-dir\") pod \"csi-hostpathplugin-vsdpm\" (UID: \"c0ac631f-5d13-4814-84d3-61f7bd704f9e\") " pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596334 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/45732d49-ce14-439b-8b7a-fb9fdf267fb1-tmpfs\") pod \"packageserver-d55dfcdfc-tg5lv\" (UID: \"45732d49-ce14-439b-8b7a-fb9fdf267fb1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596351 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c0ac631f-5d13-4814-84d3-61f7bd704f9e-mountpoint-dir\") pod \"csi-hostpathplugin-vsdpm\" (UID: \"c0ac631f-5d13-4814-84d3-61f7bd704f9e\") " pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596378 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb2d4bd5-8dac-4b9d-9125-51fca045c2b3-config\") pod \"kube-controller-manager-operator-78b949d7b-q8c9m\" (UID: \"eb2d4bd5-8dac-4b9d-9125-51fca045c2b3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q8c9m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596394 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/62823972-71dc-4f7d-b4f6-da0cb90dcf36-srv-cert\") pod \"catalog-operator-68c6474976-9c2mr\" (UID: \"62823972-71dc-4f7d-b4f6-da0cb90dcf36\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596409 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkw4k\" (UniqueName: \"kubernetes.io/projected/45732d49-ce14-439b-8b7a-fb9fdf267fb1-kube-api-access-vkw4k\") pod \"packageserver-d55dfcdfc-tg5lv\" (UID: \"45732d49-ce14-439b-8b7a-fb9fdf267fb1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596439 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7a94cb6d-599c-4dae-b7db-43f67a6620b6-signing-key\") pod \"service-ca-9c57cc56f-99q6g\" (UID: \"7a94cb6d-599c-4dae-b7db-43f67a6620b6\") " pod="openshift-service-ca/service-ca-9c57cc56f-99q6g" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596454 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c0ac631f-5d13-4814-84d3-61f7bd704f9e-plugins-dir\") pod \"csi-hostpathplugin-vsdpm\" (UID: \"c0ac631f-5d13-4814-84d3-61f7bd704f9e\") " pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596471 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7jwh\" (UniqueName: \"kubernetes.io/projected/62823972-71dc-4f7d-b4f6-da0cb90dcf36-kube-api-access-k7jwh\") pod \"catalog-operator-68c6474976-9c2mr\" (UID: \"62823972-71dc-4f7d-b4f6-da0cb90dcf36\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596489 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qwml\" (UniqueName: \"kubernetes.io/projected/055ead87-3e57-4053-a026-24566879437c-kube-api-access-2qwml\") pod \"collect-profiles-29421225-96fgb\" (UID: \"055ead87-3e57-4053-a026-24566879437c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421225-96fgb" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596513 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlgpp\" (UniqueName: \"kubernetes.io/projected/6c10edf9-cf75-4633-8d4c-9bbb9d9b6339-kube-api-access-mlgpp\") pod \"package-server-manager-789f6589d5-k8bgl\" (UID: \"6c10edf9-cf75-4633-8d4c-9bbb9d9b6339\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596539 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c0ac631f-5d13-4814-84d3-61f7bd704f9e-csi-data-dir\") pod \"csi-hostpathplugin-vsdpm\" (UID: \"c0ac631f-5d13-4814-84d3-61f7bd704f9e\") " pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596555 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8q5s\" (UniqueName: \"kubernetes.io/projected/67324e78-0fdb-4c67-a755-e2277801c106-kube-api-access-k8q5s\") pod \"multus-admission-controller-857f4d67dd-mnksj\" (UID: \"67324e78-0fdb-4c67-a755-e2277801c106\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mnksj" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596599 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c0ac631f-5d13-4814-84d3-61f7bd704f9e-socket-dir\") pod \"csi-hostpathplugin-vsdpm\" (UID: \"c0ac631f-5d13-4814-84d3-61f7bd704f9e\") " pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596617 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/648e0288-ccba-422a-a8c8-b41d12a20845-serving-cert\") pod \"service-ca-operator-777779d784-pk58m\" (UID: \"648e0288-ccba-422a-a8c8-b41d12a20845\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pk58m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596634 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/45732d49-ce14-439b-8b7a-fb9fdf267fb1-webhook-cert\") pod \"packageserver-d55dfcdfc-tg5lv\" (UID: \"45732d49-ce14-439b-8b7a-fb9fdf267fb1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596667 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/43d3560f-ad7b-4037-a980-0ed995e7e1d2-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9jtv8\" (UID: \"43d3560f-ad7b-4037-a980-0ed995e7e1d2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9jtv8" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596687 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/dac1ac9b-4ddd-41b2-813b-ed7d8a71642f-default-certificate\") pod \"router-default-5444994796-mj7jx\" (UID: \"dac1ac9b-4ddd-41b2-813b-ed7d8a71642f\") " pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596703 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb2d4bd5-8dac-4b9d-9125-51fca045c2b3-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-q8c9m\" (UID: \"eb2d4bd5-8dac-4b9d-9125-51fca045c2b3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q8c9m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596721 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/401617ee-8306-4176-9af1-4ed2a5f9af0d-cert\") pod \"ingress-canary-c9nzk\" (UID: \"401617ee-8306-4176-9af1-4ed2a5f9af0d\") " pod="openshift-ingress-canary/ingress-canary-c9nzk" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596736 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2wk8\" (UniqueName: \"kubernetes.io/projected/54818ac0-7b5e-452f-b2f7-6a87a667a76f-kube-api-access-n2wk8\") pod \"marketplace-operator-79b997595-kv8sn\" (UID: \"54818ac0-7b5e-452f-b2f7-6a87a667a76f\") " pod="openshift-marketplace/marketplace-operator-79b997595-kv8sn" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596751 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfjh9\" (UniqueName: \"kubernetes.io/projected/2474e03a-283f-48af-ac3a-eebe08168b8f-kube-api-access-xfjh9\") pod \"dns-default-zjjp6\" (UID: \"2474e03a-283f-48af-ac3a-eebe08168b8f\") " pod="openshift-dns/dns-default-zjjp6" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.596829 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/648e0288-ccba-422a-a8c8-b41d12a20845-config\") pod \"service-ca-operator-777779d784-pk58m\" (UID: \"648e0288-ccba-422a-a8c8-b41d12a20845\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pk58m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.597427 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/564831de-2ff3-4051-a1a0-c5317bf10ca1-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gmk4v\" (UID: \"564831de-2ff3-4051-a1a0-c5317bf10ca1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmk4v" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.598964 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c0ac631f-5d13-4814-84d3-61f7bd704f9e-plugins-dir\") pod \"csi-hostpathplugin-vsdpm\" (UID: \"c0ac631f-5d13-4814-84d3-61f7bd704f9e\") " pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.599421 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dac1ac9b-4ddd-41b2-813b-ed7d8a71642f-service-ca-bundle\") pod \"router-default-5444994796-mj7jx\" (UID: \"dac1ac9b-4ddd-41b2-813b-ed7d8a71642f\") " pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.599494 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c0ac631f-5d13-4814-84d3-61f7bd704f9e-registration-dir\") pod \"csi-hostpathplugin-vsdpm\" (UID: \"c0ac631f-5d13-4814-84d3-61f7bd704f9e\") " pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.599947 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncvbc\" (UniqueName: \"kubernetes.io/projected/7b26f121-04f6-4501-84b2-1833b927aa14-kube-api-access-ncvbc\") pod \"openshift-config-operator-7777fb866f-j4zxt\" (UID: \"7b26f121-04f6-4501-84b2-1833b927aa14\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.600079 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/45732d49-ce14-439b-8b7a-fb9fdf267fb1-tmpfs\") pod \"packageserver-d55dfcdfc-tg5lv\" (UID: \"45732d49-ce14-439b-8b7a-fb9fdf267fb1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.600141 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c0ac631f-5d13-4814-84d3-61f7bd704f9e-mountpoint-dir\") pod \"csi-hostpathplugin-vsdpm\" (UID: \"c0ac631f-5d13-4814-84d3-61f7bd704f9e\") " pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.600326 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2474e03a-283f-48af-ac3a-eebe08168b8f-config-volume\") pod \"dns-default-zjjp6\" (UID: \"2474e03a-283f-48af-ac3a-eebe08168b8f\") " pod="openshift-dns/dns-default-zjjp6" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.600834 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb2d4bd5-8dac-4b9d-9125-51fca045c2b3-config\") pod \"kube-controller-manager-operator-78b949d7b-q8c9m\" (UID: \"eb2d4bd5-8dac-4b9d-9125-51fca045c2b3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q8c9m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.604739 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c0ac631f-5d13-4814-84d3-61f7bd704f9e-csi-data-dir\") pod \"csi-hostpathplugin-vsdpm\" (UID: \"c0ac631f-5d13-4814-84d3-61f7bd704f9e\") " pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.604921 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c0ac631f-5d13-4814-84d3-61f7bd704f9e-socket-dir\") pod \"csi-hostpathplugin-vsdpm\" (UID: \"c0ac631f-5d13-4814-84d3-61f7bd704f9e\") " pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.608222 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/6c10edf9-cf75-4633-8d4c-9bbb9d9b6339-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-k8bgl\" (UID: \"6c10edf9-cf75-4633-8d4c-9bbb9d9b6339\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.608573 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/54818ac0-7b5e-452f-b2f7-6a87a667a76f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kv8sn\" (UID: \"54818ac0-7b5e-452f-b2f7-6a87a667a76f\") " pod="openshift-marketplace/marketplace-operator-79b997595-kv8sn" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.609799 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a7c02c2b-70ae-4b6f-85ae-5a38b737a73b-srv-cert\") pod \"olm-operator-6b444d44fb-c2dfm\" (UID: \"a7c02c2b-70ae-4b6f-85ae-5a38b737a73b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c2dfm" Dec 09 09:49:42 crc kubenswrapper[4824]: E1209 09:49:42.610067 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:43.110047817 +0000 UTC m=+139.444552484 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.610413 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/055ead87-3e57-4053-a026-24566879437c-config-volume\") pod \"collect-profiles-29421225-96fgb\" (UID: \"055ead87-3e57-4053-a026-24566879437c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421225-96fgb" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.613269 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/648e0288-ccba-422a-a8c8-b41d12a20845-serving-cert\") pod \"service-ca-operator-777779d784-pk58m\" (UID: \"648e0288-ccba-422a-a8c8-b41d12a20845\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pk58m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.613674 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/37c820ff-b4bf-41ae-bcf3-8e0f7368bb12-proxy-tls\") pod \"machine-config-controller-84d6567774-hjg74\" (UID: \"37c820ff-b4bf-41ae-bcf3-8e0f7368bb12\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hjg74" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.615961 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6d910d95-4961-4d28-9599-087153ce7d2f-bound-sa-token\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.616525 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/055ead87-3e57-4053-a026-24566879437c-secret-volume\") pod \"collect-profiles-29421225-96fgb\" (UID: \"055ead87-3e57-4053-a026-24566879437c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421225-96fgb" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.627391 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" event={"ID":"cd6c14a0-99d8-485d-81f3-cc4ebe5a943b","Type":"ContainerStarted","Data":"97c7e3b1f76efb699d4015e9514f819e32272fc0f7184743e1eb1a7cea87380e"} Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.627451 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" event={"ID":"cd6c14a0-99d8-485d-81f3-cc4ebe5a943b","Type":"ContainerStarted","Data":"360f84057d32ba228af5aacae31bc6cca680a0c254cdc94579eaa781190e0ab2"} Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.627813 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.631497 4824 patch_prober.go:28] interesting pod/console-operator-58897d9998-wgtx2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.631573 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" podUID="cd6c14a0-99d8-485d-81f3-cc4ebe5a943b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.634526 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a7c02c2b-70ae-4b6f-85ae-5a38b737a73b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-c2dfm\" (UID: \"a7c02c2b-70ae-4b6f-85ae-5a38b737a73b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c2dfm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.634676 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/dac1ac9b-4ddd-41b2-813b-ed7d8a71642f-default-certificate\") pod \"router-default-5444994796-mj7jx\" (UID: \"dac1ac9b-4ddd-41b2-813b-ed7d8a71642f\") " pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.640716 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/62823972-71dc-4f7d-b4f6-da0cb90dcf36-srv-cert\") pod \"catalog-operator-68c6474976-9c2mr\" (UID: \"62823972-71dc-4f7d-b4f6-da0cb90dcf36\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.640878 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5d1a49e0-07ed-4b75-8ce1-f6c8e9930312-node-bootstrap-token\") pod \"machine-config-server-g2cbw\" (UID: \"5d1a49e0-07ed-4b75-8ce1-f6c8e9930312\") " pod="openshift-machine-config-operator/machine-config-server-g2cbw" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.641913 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/62823972-71dc-4f7d-b4f6-da0cb90dcf36-profile-collector-cert\") pod \"catalog-operator-68c6474976-9c2mr\" (UID: \"62823972-71dc-4f7d-b4f6-da0cb90dcf36\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.642329 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb2d4bd5-8dac-4b9d-9125-51fca045c2b3-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-q8c9m\" (UID: \"eb2d4bd5-8dac-4b9d-9125-51fca045c2b3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q8c9m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.642897 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2474e03a-283f-48af-ac3a-eebe08168b8f-metrics-tls\") pod \"dns-default-zjjp6\" (UID: \"2474e03a-283f-48af-ac3a-eebe08168b8f\") " pod="openshift-dns/dns-default-zjjp6" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.643827 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/67324e78-0fdb-4c67-a755-e2277801c106-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-mnksj\" (UID: \"67324e78-0fdb-4c67-a755-e2277801c106\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mnksj" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.643981 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7a94cb6d-599c-4dae-b7db-43f67a6620b6-signing-cabundle\") pod \"service-ca-9c57cc56f-99q6g\" (UID: \"7a94cb6d-599c-4dae-b7db-43f67a6620b6\") " pod="openshift-service-ca/service-ca-9c57cc56f-99q6g" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.645734 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab5c6701-8122-4d9e-829d-93018eb67d7b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-pvgjx\" (UID: \"ab5c6701-8122-4d9e-829d-93018eb67d7b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pvgjx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.646167 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dac1ac9b-4ddd-41b2-813b-ed7d8a71642f-metrics-certs\") pod \"router-default-5444994796-mj7jx\" (UID: \"dac1ac9b-4ddd-41b2-813b-ed7d8a71642f\") " pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.646514 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv" event={"ID":"2d573cb8-34fa-40ac-98e6-c4f18caf4e0b","Type":"ContainerStarted","Data":"9a26e71486ae07234e59f7f88717ebf731f51d54029951bf6107b3239eccb2a6"} Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.646564 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv" event={"ID":"2d573cb8-34fa-40ac-98e6-c4f18caf4e0b","Type":"ContainerStarted","Data":"dbccaeda174365dcbd986aed571d82d3572d104f23bc091f22eb0d456a26677c"} Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.647616 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/564831de-2ff3-4051-a1a0-c5317bf10ca1-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gmk4v\" (UID: \"564831de-2ff3-4051-a1a0-c5317bf10ca1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmk4v" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.649735 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/45732d49-ce14-439b-8b7a-fb9fdf267fb1-webhook-cert\") pod \"packageserver-d55dfcdfc-tg5lv\" (UID: \"45732d49-ce14-439b-8b7a-fb9fdf267fb1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.649988 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/43d3560f-ad7b-4037-a980-0ed995e7e1d2-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9jtv8\" (UID: \"43d3560f-ad7b-4037-a980-0ed995e7e1d2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9jtv8" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.650155 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7a94cb6d-599c-4dae-b7db-43f67a6620b6-signing-key\") pod \"service-ca-9c57cc56f-99q6g\" (UID: \"7a94cb6d-599c-4dae-b7db-43f67a6620b6\") " pod="openshift-service-ca/service-ca-9c57cc56f-99q6g" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.651767 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/dac1ac9b-4ddd-41b2-813b-ed7d8a71642f-stats-auth\") pod \"router-default-5444994796-mj7jx\" (UID: \"dac1ac9b-4ddd-41b2-813b-ed7d8a71642f\") " pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.652039 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5d1a49e0-07ed-4b75-8ce1-f6c8e9930312-certs\") pod \"machine-config-server-g2cbw\" (UID: \"5d1a49e0-07ed-4b75-8ce1-f6c8e9930312\") " pod="openshift-machine-config-operator/machine-config-server-g2cbw" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.656732 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/401617ee-8306-4176-9af1-4ed2a5f9af0d-cert\") pod \"ingress-canary-c9nzk\" (UID: \"401617ee-8306-4176-9af1-4ed2a5f9af0d\") " pod="openshift-ingress-canary/ingress-canary-c9nzk" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.657274 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" event={"ID":"a7eb77ab-add0-4601-a279-3f8fc956b52f","Type":"ContainerStarted","Data":"6da487162b1512cf0a7f19ec366feec1802e6cb6f9163c7b9a8b00823c0c4046"} Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.658503 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/54818ac0-7b5e-452f-b2f7-6a87a667a76f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kv8sn\" (UID: \"54818ac0-7b5e-452f-b2f7-6a87a667a76f\") " pod="openshift-marketplace/marketplace-operator-79b997595-kv8sn" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.660154 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d6vn\" (UniqueName: \"kubernetes.io/projected/a4e4cf9c-48df-43b1-8ad7-9826c6220a45-kube-api-access-5d6vn\") pod \"machine-approver-56656f9798-mjsbv\" (UID: \"a4e4cf9c-48df-43b1-8ad7-9826c6220a45\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjsbv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.660826 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/45732d49-ce14-439b-8b7a-fb9fdf267fb1-apiservice-cert\") pod \"packageserver-d55dfcdfc-tg5lv\" (UID: \"45732d49-ce14-439b-8b7a-fb9fdf267fb1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.661546 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" event={"ID":"b18b84da-1277-4a24-b434-c93eb83a3af8","Type":"ContainerStarted","Data":"411a09bdb1a19b07848809653201eee2878a638679bbb1821a4de0a6cf2e9e89"} Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.662626 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qr8wb" event={"ID":"66cd34f5-c1e9-4b01-80ba-71b574b97ced","Type":"ContainerStarted","Data":"f2dc1a288340b39d996f0a5c70f9b7133c87a8b9306b1dac2cfabed254252e5d"} Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.667738 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" event={"ID":"f93c02c2-b12a-489d-9f82-1ca3dadd18d5","Type":"ContainerStarted","Data":"8b629f543cea3ce2cca3ac8fd3f64a6fe7e440be159f7bc8a4860bc59e4a9a87"} Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.669175 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" event={"ID":"c3400b7c-5666-4b36-a50f-7800f6737527","Type":"ContainerStarted","Data":"dad68de7a3715e9251b47e57ae7fb2e54378e126e3434ff471ccba56a86ab22c"} Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.670829 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zv6dz" event={"ID":"76dddf3f-2b2c-4644-ab68-402fa1e6b4d3","Type":"ContainerStarted","Data":"8bef62b6e1940e6f43d153b72360f92d1dca7a8d07126c0e76659fb6ce98d6e7"} Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.678756 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.684638 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/06fc7061-cbbe-459e-bcde-09d1c700ceba-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-847gg\" (UID: \"06fc7061-cbbe-459e-bcde-09d1c700ceba\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-847gg" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.695070 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bfj4\" (UniqueName: \"kubernetes.io/projected/14ebfe85-ea7c-49a4-a2e2-f8822c3a111e-kube-api-access-5bfj4\") pod \"machine-api-operator-5694c8668f-4rv7m\" (UID: \"14ebfe85-ea7c-49a4-a2e2-f8822c3a111e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4rv7m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.698116 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:42 crc kubenswrapper[4824]: E1209 09:49:42.698595 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:43.198567673 +0000 UTC m=+139.533072340 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.698709 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: E1209 09:49:42.699229 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:43.199209105 +0000 UTC m=+139.533713772 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.733059 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfrn2\" (UniqueName: \"kubernetes.io/projected/06fc7061-cbbe-459e-bcde-09d1c700ceba-kube-api-access-dfrn2\") pod \"cluster-image-registry-operator-dc59b4c8b-847gg\" (UID: \"06fc7061-cbbe-459e-bcde-09d1c700ceba\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-847gg" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.745803 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfjh9\" (UniqueName: \"kubernetes.io/projected/2474e03a-283f-48af-ac3a-eebe08168b8f-kube-api-access-xfjh9\") pod \"dns-default-zjjp6\" (UID: \"2474e03a-283f-48af-ac3a-eebe08168b8f\") " pod="openshift-dns/dns-default-zjjp6" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.752186 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhjpz\" (UniqueName: \"kubernetes.io/projected/a7c02c2b-70ae-4b6f-85ae-5a38b737a73b-kube-api-access-vhjpz\") pod \"olm-operator-6b444d44fb-c2dfm\" (UID: \"a7c02c2b-70ae-4b6f-85ae-5a38b737a73b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c2dfm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.757679 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjsbv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.776662 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxqf8\" (UniqueName: \"kubernetes.io/projected/dac1ac9b-4ddd-41b2-813b-ed7d8a71642f-kube-api-access-fxqf8\") pod \"router-default-5444994796-mj7jx\" (UID: \"dac1ac9b-4ddd-41b2-813b-ed7d8a71642f\") " pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 09:49:42 crc kubenswrapper[4824]: W1209 09:49:42.776872 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4e4cf9c_48df_43b1_8ad7_9826c6220a45.slice/crio-f819a4b926426400cd6f7817de6d84faf317462dd004426f1d70a57a761243e6 WatchSource:0}: Error finding container f819a4b926426400cd6f7817de6d84faf317462dd004426f1d70a57a761243e6: Status 404 returned error can't find the container with id f819a4b926426400cd6f7817de6d84faf317462dd004426f1d70a57a761243e6 Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.795349 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-4rv7m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.796574 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-c6559" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.800349 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:42 crc kubenswrapper[4824]: E1209 09:49:42.800900 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:43.3008714 +0000 UTC m=+139.635376067 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.806214 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hglrf\" (UniqueName: \"kubernetes.io/projected/648e0288-ccba-422a-a8c8-b41d12a20845-kube-api-access-hglrf\") pod \"service-ca-operator-777779d784-pk58m\" (UID: \"648e0288-ccba-422a-a8c8-b41d12a20845\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-pk58m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.813514 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.820043 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-b9bnh" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.834552 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.837758 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-jgwrj"] Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.858066 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7b9dd"] Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.858122 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2t6n7"] Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.866639 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg7lm\" (UniqueName: \"kubernetes.io/projected/c0ac631f-5d13-4814-84d3-61f7bd704f9e-kube-api-access-qg7lm\") pod \"csi-hostpathplugin-vsdpm\" (UID: \"c0ac631f-5d13-4814-84d3-61f7bd704f9e\") " pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.879444 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-k4fpg"] Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.887666 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbsxw\" (UniqueName: \"kubernetes.io/projected/43d3560f-ad7b-4037-a980-0ed995e7e1d2-kube-api-access-qbsxw\") pod \"control-plane-machine-set-operator-78cbb6b69f-9jtv8\" (UID: \"43d3560f-ad7b-4037-a980-0ed995e7e1d2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9jtv8" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.899698 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkw4k\" (UniqueName: \"kubernetes.io/projected/45732d49-ce14-439b-8b7a-fb9fdf267fb1-kube-api-access-vkw4k\") pod \"packageserver-d55dfcdfc-tg5lv\" (UID: \"45732d49-ce14-439b-8b7a-fb9fdf267fb1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.901997 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:42 crc kubenswrapper[4824]: E1209 09:49:42.902543 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:43.402527527 +0000 UTC m=+139.737032194 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.909853 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-ldw7r"] Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.917626 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7jwh\" (UniqueName: \"kubernetes.io/projected/62823972-71dc-4f7d-b4f6-da0cb90dcf36-kube-api-access-k7jwh\") pod \"catalog-operator-68c6474976-9c2mr\" (UID: \"62823972-71dc-4f7d-b4f6-da0cb90dcf36\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.932036 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9jtv8" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.938643 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qwml\" (UniqueName: \"kubernetes.io/projected/055ead87-3e57-4053-a026-24566879437c-kube-api-access-2qwml\") pod \"collect-profiles-29421225-96fgb\" (UID: \"055ead87-3e57-4053-a026-24566879437c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421225-96fgb" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.948562 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.950991 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/564831de-2ff3-4051-a1a0-c5317bf10ca1-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gmk4v\" (UID: \"564831de-2ff3-4051-a1a0-c5317bf10ca1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmk4v" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.952203 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlgpp\" (UniqueName: \"kubernetes.io/projected/6c10edf9-cf75-4633-8d4c-9bbb9d9b6339-kube-api-access-mlgpp\") pod \"package-server-manager-789f6589d5-k8bgl\" (UID: \"6c10edf9-cf75-4633-8d4c-9bbb9d9b6339\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.956002 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lv67\" (UniqueName: \"kubernetes.io/projected/401617ee-8306-4176-9af1-4ed2a5f9af0d-kube-api-access-8lv67\") pod \"ingress-canary-c9nzk\" (UID: \"401617ee-8306-4176-9af1-4ed2a5f9af0d\") " pod="openshift-ingress-canary/ingress-canary-c9nzk" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.963259 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.970702 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-pk58m" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.972646 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-847gg" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.976256 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.976961 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8q5s\" (UniqueName: \"kubernetes.io/projected/67324e78-0fdb-4c67-a755-e2277801c106-kube-api-access-k8q5s\") pod \"multus-admission-controller-857f4d67dd-mnksj\" (UID: \"67324e78-0fdb-4c67-a755-e2277801c106\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mnksj" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.988250 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x7j5h"] Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.988640 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c2dfm" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.994453 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjlmb\" (UniqueName: \"kubernetes.io/projected/ab5c6701-8122-4d9e-829d-93018eb67d7b-kube-api-access-hjlmb\") pod \"openshift-controller-manager-operator-756b6f6bc6-pvgjx\" (UID: \"ab5c6701-8122-4d9e-829d-93018eb67d7b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pvgjx" Dec 09 09:49:42 crc kubenswrapper[4824]: I1209 09:49:42.994943 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-mnksj" Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.002035 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-c9nzk" Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.002995 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:43 crc kubenswrapper[4824]: E1209 09:49:43.003238 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:43.503203501 +0000 UTC m=+139.837708178 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.003379 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:43 crc kubenswrapper[4824]: E1209 09:49:43.003937 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:43.503926715 +0000 UTC m=+139.838431382 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.009284 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-zjjp6" Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.020893 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt"] Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.021416 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb2d4bd5-8dac-4b9d-9125-51fca045c2b3-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-q8c9m\" (UID: \"eb2d4bd5-8dac-4b9d-9125-51fca045c2b3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q8c9m" Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.030156 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.035248 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2wk8\" (UniqueName: \"kubernetes.io/projected/54818ac0-7b5e-452f-b2f7-6a87a667a76f-kube-api-access-n2wk8\") pod \"marketplace-operator-79b997595-kv8sn\" (UID: \"54818ac0-7b5e-452f-b2f7-6a87a667a76f\") " pod="openshift-marketplace/marketplace-operator-79b997595-kv8sn" Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.059200 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgqnn\" (UniqueName: \"kubernetes.io/projected/5d1a49e0-07ed-4b75-8ce1-f6c8e9930312-kube-api-access-mgqnn\") pod \"machine-config-server-g2cbw\" (UID: \"5d1a49e0-07ed-4b75-8ce1-f6c8e9930312\") " pod="openshift-machine-config-operator/machine-config-server-g2cbw" Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.079609 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-td5l8\" (UniqueName: \"kubernetes.io/projected/7a94cb6d-599c-4dae-b7db-43f67a6620b6-kube-api-access-td5l8\") pod \"service-ca-9c57cc56f-99q6g\" (UID: \"7a94cb6d-599c-4dae-b7db-43f67a6620b6\") " pod="openshift-service-ca/service-ca-9c57cc56f-99q6g" Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.090850 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfvr4\" (UniqueName: \"kubernetes.io/projected/0f899ebd-d4e0-4e7c-82ea-914abad9dab6-kube-api-access-tfvr4\") pod \"migrator-59844c95c7-zl4mr\" (UID: \"0f899ebd-d4e0-4e7c-82ea-914abad9dab6\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zl4mr" Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.090890 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-c6559"] Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.091968 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4rv7m"] Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.104422 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:43 crc kubenswrapper[4824]: E1209 09:49:43.104632 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:43.604594639 +0000 UTC m=+139.939099306 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.104809 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:43 crc kubenswrapper[4824]: E1209 09:49:43.105218 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:43.60520621 +0000 UTC m=+139.939710937 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.113617 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g2fd\" (UniqueName: \"kubernetes.io/projected/37c820ff-b4bf-41ae-bcf3-8e0f7368bb12-kube-api-access-6g2fd\") pod \"machine-config-controller-84d6567774-hjg74\" (UID: \"37c820ff-b4bf-41ae-bcf3-8e0f7368bb12\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hjg74" Dec 09 09:49:43 crc kubenswrapper[4824]: W1209 09:49:43.130239 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9966a4dc_9922_434f_8f3e_6efc5db8445e.slice/crio-ce37527facadee8b8b98895624ee042891add23039107fc5ea462563c265b2c6 WatchSource:0}: Error finding container ce37527facadee8b8b98895624ee042891add23039107fc5ea462563c265b2c6: Status 404 returned error can't find the container with id ce37527facadee8b8b98895624ee042891add23039107fc5ea462563c265b2c6 Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.160087 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q8c9m" Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.186245 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pvgjx" Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.195002 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zl4mr" Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.201861 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmk4v" Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.212500 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421225-96fgb" Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.213291 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:43 crc kubenswrapper[4824]: E1209 09:49:43.213850 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:43.713825024 +0000 UTC m=+140.048329691 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.221052 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kv8sn" Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.224561 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-g2cbw" Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.247980 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-99q6g" Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.258897 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hjg74" Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.315202 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:43 crc kubenswrapper[4824]: E1209 09:49:43.315620 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:43.815605695 +0000 UTC m=+140.150110362 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.416016 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:43 crc kubenswrapper[4824]: E1209 09:49:43.416221 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:43.916186466 +0000 UTC m=+140.250691143 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.416317 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:43 crc kubenswrapper[4824]: E1209 09:49:43.416721 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:43.916703542 +0000 UTC m=+140.251208209 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.522757 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:43 crc kubenswrapper[4824]: E1209 09:49:43.524656 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:44.024619394 +0000 UTC m=+140.359124061 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.525440 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:43 crc kubenswrapper[4824]: E1209 09:49:43.526297 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:44.026266897 +0000 UTC m=+140.360771564 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.631168 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:43 crc kubenswrapper[4824]: E1209 09:49:43.631367 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:44.131325045 +0000 UTC m=+140.465829712 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.631620 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:43 crc kubenswrapper[4824]: E1209 09:49:43.632128 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:44.132120032 +0000 UTC m=+140.466624689 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.709224 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qr8wb" event={"ID":"66cd34f5-c1e9-4b01-80ba-71b574b97ced","Type":"ContainerStarted","Data":"1946a956fe191d80b8a2a6aae3ece80b79997e126d7c0b54ddd447a33208bf50"} Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.714124 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-7b9dd" event={"ID":"f4616db0-42ba-4a3b-88f2-57a0663d0a15","Type":"ContainerStarted","Data":"b9e642826791faecf1c9c3b7db12e9e60a1786fb10c7df8e10ab97515f5635ee"} Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.725073 4824 generic.go:334] "Generic (PLEG): container finished" podID="f93c02c2-b12a-489d-9f82-1ca3dadd18d5" containerID="16d1cb3ecc3622d4879a7a688522964fe6ad759d4556b14bb3030ac08e6646fc" exitCode=0 Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.725151 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" event={"ID":"f93c02c2-b12a-489d-9f82-1ca3dadd18d5","Type":"ContainerDied","Data":"16d1cb3ecc3622d4879a7a688522964fe6ad759d4556b14bb3030ac08e6646fc"} Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.732614 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:43 crc kubenswrapper[4824]: E1209 09:49:43.732953 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:44.23293252 +0000 UTC m=+140.567437187 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.735004 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ldw7r" event={"ID":"0c5e7133-21dc-46e0-b334-9d36c0427aa3","Type":"ContainerStarted","Data":"800371374fe33dfa445ea3c80d7fb2b4d04e84086246d829bf952bc1cd47e079"} Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.745301 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-mj7jx" event={"ID":"dac1ac9b-4ddd-41b2-813b-ed7d8a71642f","Type":"ContainerStarted","Data":"d4506c76b7813807b95f701d6a6660b2fcdaac18b774ba8d8ef5b96dea985cf9"} Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.756157 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-c6559" event={"ID":"9966a4dc-9922-434f-8f3e-6efc5db8445e","Type":"ContainerStarted","Data":"ce37527facadee8b8b98895624ee042891add23039107fc5ea462563c265b2c6"} Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.779485 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k4fpg" event={"ID":"76df4be1-7bf3-461d-a9d7-6a2be7c0e425","Type":"ContainerStarted","Data":"f916c9adfaec1416358a67293be318e3b7fd37ab4272bdc07354ba712ee6dd4a"} Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.779549 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k4fpg" event={"ID":"76df4be1-7bf3-461d-a9d7-6a2be7c0e425","Type":"ContainerStarted","Data":"e4f48a7a498bdba99c3c6b7c73b947b68716079681d4dc6a933d5b30c70f0109"} Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.782213 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4rv7m" event={"ID":"14ebfe85-ea7c-49a4-a2e2-f8822c3a111e","Type":"ContainerStarted","Data":"18d6bef160fc93f29feca2884e426233b4e03d9d09bbca99c3f656829b2b97e9"} Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.793707 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" event={"ID":"b18b84da-1277-4a24-b434-c93eb83a3af8","Type":"ContainerStarted","Data":"c5836d095b038fd556cbc9efa83d716da01139ebf660b51cb119df979bc7ef95"} Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.797128 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.803213 4824 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-tqhnj container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.803279 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" podUID="b18b84da-1277-4a24-b434-c93eb83a3af8" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.807314 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjsbv" event={"ID":"a4e4cf9c-48df-43b1-8ad7-9826c6220a45","Type":"ContainerStarted","Data":"c39f8664dd85b49545870143650fa6ee2293f70c01b81b3feda899f76b392564"} Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.807359 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjsbv" event={"ID":"a4e4cf9c-48df-43b1-8ad7-9826c6220a45","Type":"ContainerStarted","Data":"f819a4b926426400cd6f7817de6d84faf317462dd004426f1d70a57a761243e6"} Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.816831 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2t6n7" event={"ID":"6042da41-13ef-460e-aee4-491c7758f635","Type":"ContainerStarted","Data":"53df42f98eaa592904766b544a7ed82644ad0871279d2f9aad9907a849df3468"} Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.829367 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" event={"ID":"7b26f121-04f6-4501-84b2-1833b927aa14","Type":"ContainerStarted","Data":"bf3960efca266cd90ba0931ed81dc510feb2908b7c2e0dacb10642e3b63d2867"} Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.837542 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:43 crc kubenswrapper[4824]: E1209 09:49:43.840388 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:44.340372946 +0000 UTC m=+140.674877683 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.848069 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jgwrj" event={"ID":"00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53","Type":"ContainerStarted","Data":"b6b6179c37a9a43331580d39efdf7229a42f58c45e1b7363a8214dfcda39a846"} Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.885177 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" event={"ID":"c3400b7c-5666-4b36-a50f-7800f6737527","Type":"ContainerStarted","Data":"ededff8a2f4ee4e72db8028a40598edd0ce0981b53e5b3d8e13b10053bf64cd1"} Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.937675 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zv6dz" event={"ID":"76dddf3f-2b2c-4644-ab68-402fa1e6b4d3","Type":"ContainerStarted","Data":"7e7d0de41d1d6aafaf0668cc4c71554de02d73031c22bdb94a8782b307cdec97"} Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.960760 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x7j5h" event={"ID":"5047f85e-0349-40fc-b966-1956bd66cbbc","Type":"ContainerStarted","Data":"68073cfe791e2a960db1b357ed9554b8e60a0cb262f94029306f44b7a9a1d451"} Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.960886 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x7j5h" event={"ID":"5047f85e-0349-40fc-b966-1956bd66cbbc","Type":"ContainerStarted","Data":"0b126a5dc3eb19c8181478a1e00b1116dabed33cce508135af0c95841d866e10"} Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.989487 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" event={"ID":"a7eb77ab-add0-4601-a279-3f8fc956b52f","Type":"ContainerStarted","Data":"baba12e8836c2703824b250129d4413aa64c3efe56d42e645c1bb5b413d11880"} Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.989897 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv" Dec 09 09:49:43 crc kubenswrapper[4824]: I1209 09:49:43.990386 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:44 crc kubenswrapper[4824]: I1209 09:49:44.008321 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:44 crc kubenswrapper[4824]: E1209 09:49:44.010243 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:44.510221463 +0000 UTC m=+140.844726130 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:44 crc kubenswrapper[4824]: I1209 09:49:44.024230 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" Dec 09 09:49:44 crc kubenswrapper[4824]: I1209 09:49:44.090768 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv" Dec 09 09:49:44 crc kubenswrapper[4824]: I1209 09:49:44.114268 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:44 crc kubenswrapper[4824]: E1209 09:49:44.122170 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:44.622133016 +0000 UTC m=+140.956637703 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:44 crc kubenswrapper[4824]: I1209 09:49:44.193578 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv"] Dec 09 09:49:44 crc kubenswrapper[4824]: I1209 09:49:44.218934 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pd82d"] Dec 09 09:49:44 crc kubenswrapper[4824]: I1209 09:49:44.220059 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:44 crc kubenswrapper[4824]: E1209 09:49:44.220520 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:44.720498644 +0000 UTC m=+141.055003311 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:44 crc kubenswrapper[4824]: I1209 09:49:44.223119 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv" podStartSLOduration=120.223084338 podStartE2EDuration="2m0.223084338s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:44.179517813 +0000 UTC m=+140.514022490" watchObservedRunningTime="2025-12-09 09:49:44.223084338 +0000 UTC m=+140.557589005" Dec 09 09:49:44 crc kubenswrapper[4824]: I1209 09:49:44.226834 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" podStartSLOduration=120.226807311 podStartE2EDuration="2m0.226807311s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:44.206828847 +0000 UTC m=+140.541333514" watchObservedRunningTime="2025-12-09 09:49:44.226807311 +0000 UTC m=+140.561311998" Dec 09 09:49:44 crc kubenswrapper[4824]: I1209 09:49:44.271820 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x7j5h" podStartSLOduration=120.271801563 podStartE2EDuration="2m0.271801563s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:44.271004707 +0000 UTC m=+140.605509384" watchObservedRunningTime="2025-12-09 09:49:44.271801563 +0000 UTC m=+140.606306240" Dec 09 09:49:44 crc kubenswrapper[4824]: I1209 09:49:44.314888 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" podStartSLOduration=120.314869743 podStartE2EDuration="2m0.314869743s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:44.311910566 +0000 UTC m=+140.646415253" watchObservedRunningTime="2025-12-09 09:49:44.314869743 +0000 UTC m=+140.649374420" Dec 09 09:49:44 crc kubenswrapper[4824]: I1209 09:49:44.322913 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:44 crc kubenswrapper[4824]: E1209 09:49:44.323360 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:44.82334499 +0000 UTC m=+141.157849657 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:44 crc kubenswrapper[4824]: I1209 09:49:44.427722 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:44 crc kubenswrapper[4824]: E1209 09:49:44.428135 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:44.928112748 +0000 UTC m=+141.262617415 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:44 crc kubenswrapper[4824]: I1209 09:49:44.428753 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" podStartSLOduration=120.428740008 podStartE2EDuration="2m0.428740008s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:44.33560511 +0000 UTC m=+140.670109797" watchObservedRunningTime="2025-12-09 09:49:44.428740008 +0000 UTC m=+140.763244685" Dec 09 09:49:44 crc kubenswrapper[4824]: I1209 09:49:44.499385 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zv6dz" podStartSLOduration=120.49936175 podStartE2EDuration="2m0.49936175s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:44.45996561 +0000 UTC m=+140.794470287" watchObservedRunningTime="2025-12-09 09:49:44.49936175 +0000 UTC m=+140.833866417" Dec 09 09:49:44 crc kubenswrapper[4824]: I1209 09:49:44.530467 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:44 crc kubenswrapper[4824]: E1209 09:49:44.531878 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:45.031860293 +0000 UTC m=+141.366364960 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:44 crc kubenswrapper[4824]: I1209 09:49:44.572100 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" podStartSLOduration=120.572059318 podStartE2EDuration="2m0.572059318s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:44.536006499 +0000 UTC m=+140.870511166" watchObservedRunningTime="2025-12-09 09:49:44.572059318 +0000 UTC m=+140.906563985" Dec 09 09:49:44 crc kubenswrapper[4824]: I1209 09:49:44.585902 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:49:44 crc kubenswrapper[4824]: I1209 09:49:44.633012 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:44 crc kubenswrapper[4824]: E1209 09:49:44.633454 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:45.133422586 +0000 UTC m=+141.467927253 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:44 crc kubenswrapper[4824]: I1209 09:49:44.736698 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:44 crc kubenswrapper[4824]: E1209 09:49:44.737102 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:45.237085598 +0000 UTC m=+141.571590265 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:44 crc kubenswrapper[4824]: I1209 09:49:44.840833 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:44 crc kubenswrapper[4824]: E1209 09:49:44.841608 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:45.341585228 +0000 UTC m=+141.676089895 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:44 crc kubenswrapper[4824]: I1209 09:49:44.947588 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:44 crc kubenswrapper[4824]: E1209 09:49:44.948031 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:45.448011451 +0000 UTC m=+141.782516118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.085436 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:45 crc kubenswrapper[4824]: E1209 09:49:45.087142 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:45.587103752 +0000 UTC m=+141.921608429 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.090884 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qr8wb" event={"ID":"66cd34f5-c1e9-4b01-80ba-71b574b97ced","Type":"ContainerStarted","Data":"130554b6c79e39afa34c63867cc32b02d20edc012c37b0017179798d87ec5122"} Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.110431 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ldw7r" event={"ID":"0c5e7133-21dc-46e0-b334-9d36c0427aa3","Type":"ContainerStarted","Data":"5156e771cd0edbb835f34d16e3d11ec265dc92c76c7dce49b404389fcf18d2b7"} Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.132273 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qr8wb" podStartSLOduration=121.132255419 podStartE2EDuration="2m1.132255419s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:45.131606938 +0000 UTC m=+141.466111605" watchObservedRunningTime="2025-12-09 09:49:45.132255419 +0000 UTC m=+141.466760086" Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.140385 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jgwrj" event={"ID":"00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53","Type":"ContainerStarted","Data":"0e68e4410bc450193261bc927bafcc0e937cadd91217eb35d44efe1cfb8a9e19"} Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.151134 4824 generic.go:334] "Generic (PLEG): container finished" podID="7b26f121-04f6-4501-84b2-1833b927aa14" containerID="d35a5f26e0b6241350e0396f72787e1ddaf08e0d297f8c7d09fa444909a66abf" exitCode=0 Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.151517 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" event={"ID":"7b26f121-04f6-4501-84b2-1833b927aa14","Type":"ContainerDied","Data":"d35a5f26e0b6241350e0396f72787e1ddaf08e0d297f8c7d09fa444909a66abf"} Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.184038 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-ldw7r" podStartSLOduration=121.184008052 podStartE2EDuration="2m1.184008052s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:45.182436811 +0000 UTC m=+141.516941498" watchObservedRunningTime="2025-12-09 09:49:45.184008052 +0000 UTC m=+141.518512709" Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.197045 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:45 crc kubenswrapper[4824]: E1209 09:49:45.197396 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:45.69737976 +0000 UTC m=+142.031884427 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.206507 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k4fpg" event={"ID":"76df4be1-7bf3-461d-a9d7-6a2be7c0e425","Type":"ContainerStarted","Data":"c3a524a0a8e75aa9ae2046d14bc0cf8d4b8cedf703148e15f3d1c71e30000112"} Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.218236 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pd82d" event={"ID":"7bff53ba-5406-445c-afcb-3247516a1258","Type":"ContainerStarted","Data":"ac51aa90b68a1daa3a33fdaec76864d8f874876ab8da3b1e1cda3e935c7b1c90"} Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.234373 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4rv7m" event={"ID":"14ebfe85-ea7c-49a4-a2e2-f8822c3a111e","Type":"ContainerStarted","Data":"f35f4a09943742fc38af8b1862e7d53e7a205ca10af2489537d621849e8eb146"} Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.237079 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-g2cbw" event={"ID":"5d1a49e0-07ed-4b75-8ce1-f6c8e9930312","Type":"ContainerStarted","Data":"42719b061d1674085a7810e428507f77234be8ae55c5222c2f3eb3200916ee88"} Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.259188 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k4fpg" podStartSLOduration=121.259159681 podStartE2EDuration="2m1.259159681s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:45.25572524 +0000 UTC m=+141.590229907" watchObservedRunningTime="2025-12-09 09:49:45.259159681 +0000 UTC m=+141.593664338" Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.290242 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" event={"ID":"45732d49-ce14-439b-8b7a-fb9fdf267fb1","Type":"ContainerStarted","Data":"b56dea4e521197568fd7417d9f9ce3e27e4e4a01194a203346198096bd9d889a"} Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.299626 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:45 crc kubenswrapper[4824]: E1209 09:49:45.299912 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:45.799865954 +0000 UTC m=+142.134370621 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.300142 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:45 crc kubenswrapper[4824]: E1209 09:49:45.300729 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:45.800709412 +0000 UTC m=+142.135214079 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.312451 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-mj7jx" podStartSLOduration=121.312428944 podStartE2EDuration="2m1.312428944s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:45.305642593 +0000 UTC m=+141.640147270" watchObservedRunningTime="2025-12-09 09:49:45.312428944 +0000 UTC m=+141.646933611" Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.331461 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.355503 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2t6n7" podStartSLOduration=121.355482163 podStartE2EDuration="2m1.355482163s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:45.347795142 +0000 UTC m=+141.682299809" watchObservedRunningTime="2025-12-09 09:49:45.355482163 +0000 UTC m=+141.689986840" Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.405526 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:45 crc kubenswrapper[4824]: E1209 09:49:45.406213 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:45.906131991 +0000 UTC m=+142.240636658 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.408878 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:45 crc kubenswrapper[4824]: E1209 09:49:45.414947 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:45.914922099 +0000 UTC m=+142.249426956 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.510129 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:45 crc kubenswrapper[4824]: E1209 09:49:45.510879 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:46.010852617 +0000 UTC m=+142.345357284 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.515694 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-847gg"] Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.612195 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:45 crc kubenswrapper[4824]: E1209 09:49:45.612613 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:46.112594987 +0000 UTC m=+142.447099654 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.655343 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9jtv8"] Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.719698 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:45 crc kubenswrapper[4824]: E1209 09:49:45.720654 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:46.220628652 +0000 UTC m=+142.555133319 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.825208 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-zjjp6"] Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.826680 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:45 crc kubenswrapper[4824]: E1209 09:49:45.827050 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:46.327035674 +0000 UTC m=+142.661540341 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.839817 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.851796 4824 patch_prober.go:28] interesting pod/router-default-5444994796-mj7jx container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.851875 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Dec 09 09:49:45 crc kubenswrapper[4824]: W1209 09:49:45.913263 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2474e03a_283f_48af_ac3a_eebe08168b8f.slice/crio-5bcdfd038dfad62d26347b9b0f76373cb8910c5ec1ec959f6d8d646c8961dd46 WatchSource:0}: Error finding container 5bcdfd038dfad62d26347b9b0f76373cb8910c5ec1ec959f6d8d646c8961dd46: Status 404 returned error can't find the container with id 5bcdfd038dfad62d26347b9b0f76373cb8910c5ec1ec959f6d8d646c8961dd46 Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.927376 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:45 crc kubenswrapper[4824]: E1209 09:49:45.927992 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:46.427968876 +0000 UTC m=+142.762473543 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.957690 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-b9bnh"] Dec 09 09:49:45 crc kubenswrapper[4824]: I1209 09:49:45.986055 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-pk58m"] Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.039968 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:46 crc kubenswrapper[4824]: E1209 09:49:46.040653 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:46.540612242 +0000 UTC m=+142.875116959 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.145434 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:46 crc kubenswrapper[4824]: E1209 09:49:46.145771 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:46.645743263 +0000 UTC m=+142.980247930 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.157834 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c2dfm"] Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.207989 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr"] Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.229043 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-mnksj"] Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.233810 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl"] Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.247038 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.256171 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421225-96fgb"] Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.273012 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-zl4mr"] Dec 09 09:49:46 crc kubenswrapper[4824]: E1209 09:49:46.256534 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:46.756490066 +0000 UTC m=+143.090994733 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.288713 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q8c9m"] Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.290880 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-c9nzk"] Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.319694 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmk4v"] Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.377951 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:46 crc kubenswrapper[4824]: E1209 09:49:46.378942 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:46.878920242 +0000 UTC m=+143.213424909 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.425132 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-c6559" event={"ID":"9966a4dc-9922-434f-8f3e-6efc5db8445e","Type":"ContainerStarted","Data":"aed5347d7bc831858a5fba1b4955e3e30f25c3041abf32780bc4dbbfa8894e2d"} Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.425210 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-c6559" event={"ID":"9966a4dc-9922-434f-8f3e-6efc5db8445e","Type":"ContainerStarted","Data":"7dd045269e926a717b4ed37bdc0c1d14310ab68595f06e54dc89769db7454326"} Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.450613 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9jtv8" event={"ID":"43d3560f-ad7b-4037-a980-0ed995e7e1d2","Type":"ContainerStarted","Data":"878e0de1efc8b317dbbfa8dbe19257f7cbcc2856c6f490105aad1b1e418d1420"} Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.458227 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9jtv8" event={"ID":"43d3560f-ad7b-4037-a980-0ed995e7e1d2","Type":"ContainerStarted","Data":"bcacec588c33c6dec75d295fdef5a502425b1638de42eabd47c19d7d10b442fd"} Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.482863 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:46 crc kubenswrapper[4824]: E1209 09:49:46.483406 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:46.983388081 +0000 UTC m=+143.317892748 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.506065 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-c6559" podStartSLOduration=122.506037072 podStartE2EDuration="2m2.506037072s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:46.5041336 +0000 UTC m=+142.838638267" watchObservedRunningTime="2025-12-09 09:49:46.506037072 +0000 UTC m=+142.840541739" Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.517926 4824 generic.go:334] "Generic (PLEG): container finished" podID="7bff53ba-5406-445c-afcb-3247516a1258" containerID="87101974ba208bd7ffa1e279cefd9e7c6630b06ca3494bab4a3b7f659940a0b6" exitCode=0 Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.518056 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pd82d" event={"ID":"7bff53ba-5406-445c-afcb-3247516a1258","Type":"ContainerDied","Data":"87101974ba208bd7ffa1e279cefd9e7c6630b06ca3494bab4a3b7f659940a0b6"} Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.548888 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-hjg74"] Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.567858 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-vsdpm"] Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.567941 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9jtv8" podStartSLOduration=122.567908187 podStartE2EDuration="2m2.567908187s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:46.541915236 +0000 UTC m=+142.876419913" watchObservedRunningTime="2025-12-09 09:49:46.567908187 +0000 UTC m=+142.902412854" Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.597501 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:46 crc kubenswrapper[4824]: E1209 09:49:46.599405 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:47.099385967 +0000 UTC m=+143.433890634 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.642342 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" event={"ID":"45732d49-ce14-439b-8b7a-fb9fdf267fb1","Type":"ContainerStarted","Data":"8424dba8744dda5f7cdf5e3a63688e51799cdbdb07d370b2b3b2811ce0148354"} Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.644649 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.697070 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-g2cbw" event={"ID":"5d1a49e0-07ed-4b75-8ce1-f6c8e9930312","Type":"ContainerStarted","Data":"34e755266e527661578af4bf9f8e8d577db058934c1153a354eff9a09014f748"} Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.701848 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-99q6g"] Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.704340 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:46 crc kubenswrapper[4824]: E1209 09:49:46.706334 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:47.206313395 +0000 UTC m=+143.540818062 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.715205 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kv8sn"] Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.715668 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" podStartSLOduration=122.715641541 podStartE2EDuration="2m2.715641541s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:46.681639578 +0000 UTC m=+143.016144245" watchObservedRunningTime="2025-12-09 09:49:46.715641541 +0000 UTC m=+143.050146198" Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.753248 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pvgjx"] Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.753279 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-g2cbw" podStartSLOduration=7.753257322 podStartE2EDuration="7.753257322s" podCreationTimestamp="2025-12-09 09:49:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:46.751588817 +0000 UTC m=+143.086093484" watchObservedRunningTime="2025-12-09 09:49:46.753257322 +0000 UTC m=+143.087761989" Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.778312 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" event={"ID":"f93c02c2-b12a-489d-9f82-1ca3dadd18d5","Type":"ContainerStarted","Data":"7f35ea6a7cf0e86dcffb40020d5ac36efd4748cacb6c3aebe02f94960c512439"} Dec 09 09:49:46 crc kubenswrapper[4824]: W1209 09:49:46.780928 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37c820ff_b4bf_41ae_bcf3_8e0f7368bb12.slice/crio-f2584400dffcf142ec370abed9269795f98e222d57ee1a31a4164ae31b045f52 WatchSource:0}: Error finding container f2584400dffcf142ec370abed9269795f98e222d57ee1a31a4164ae31b045f52: Status 404 returned error can't find the container with id f2584400dffcf142ec370abed9269795f98e222d57ee1a31a4164ae31b045f52 Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.790555 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2t6n7" event={"ID":"6042da41-13ef-460e-aee4-491c7758f635","Type":"ContainerStarted","Data":"f869b0f38da55a121536155883d70314e86a902afb2c16e1178a13e050ed898c"} Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.806939 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:46 crc kubenswrapper[4824]: E1209 09:49:46.808132 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:47.308088536 +0000 UTC m=+143.642593203 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.836718 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-7b9dd" event={"ID":"f4616db0-42ba-4a3b-88f2-57a0663d0a15","Type":"ContainerStarted","Data":"9d197249549e40f4b14e229de50477cfc585b2bf5bcf6a97f662160e5678c63e"} Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.846376 4824 patch_prober.go:28] interesting pod/router-default-5444994796-mj7jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 09:49:46 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Dec 09 09:49:46 crc kubenswrapper[4824]: [+]process-running ok Dec 09 09:49:46 crc kubenswrapper[4824]: healthz check failed Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.846619 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.850362 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" event={"ID":"62823972-71dc-4f7d-b4f6-da0cb90dcf36","Type":"ContainerStarted","Data":"47c8e3f1b915fb623245548b6d7ebecb5f71d04772d10819b5803b0cd6d55e7c"} Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.887442 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-7b9dd" podStartSLOduration=122.887415512 podStartE2EDuration="2m2.887415512s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:46.887287918 +0000 UTC m=+143.221792585" watchObservedRunningTime="2025-12-09 09:49:46.887415512 +0000 UTC m=+143.221920189" Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.890645 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" podStartSLOduration=122.890633228 podStartE2EDuration="2m2.890633228s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:46.842189472 +0000 UTC m=+143.176694139" watchObservedRunningTime="2025-12-09 09:49:46.890633228 +0000 UTC m=+143.225137895" Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.900638 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-b9bnh" event={"ID":"6efb8245-07e9-4c31-b8ab-53ab3685e593","Type":"ContainerStarted","Data":"b4665be5134d1e2c0be3696c0199bc1dd74938a08bd2b258eaec36119d061d08"} Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.909374 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-b9bnh" Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.924747 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-b9bnh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.924849 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b9bnh" podUID="6efb8245-07e9-4c31-b8ab-53ab3685e593" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 09 09:49:46 crc kubenswrapper[4824]: I1209 09:49:46.926737 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:46 crc kubenswrapper[4824]: E1209 09:49:46.927125 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:47.427108931 +0000 UTC m=+143.761613598 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.021337 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jgwrj" event={"ID":"00965cee-6f6b-4d9b-8ed0-be9bdf3d9d53","Type":"ContainerStarted","Data":"2f266e72ee5d781063c1bbf31a48bf18bab74e0a7816130a675e6e10a0281cd9"} Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.033742 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:47 crc kubenswrapper[4824]: E1209 09:49:47.033956 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:47.533925656 +0000 UTC m=+143.868430313 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.034493 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:47 crc kubenswrapper[4824]: E1209 09:49:47.035060 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:47.535050853 +0000 UTC m=+143.869555510 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.053341 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-847gg" event={"ID":"06fc7061-cbbe-459e-bcde-09d1c700ceba","Type":"ContainerStarted","Data":"f074a81821d283ce5909cd88d861f94fd8234db04c7beb93b7979a15b7c13e7d"} Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.053410 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-847gg" event={"ID":"06fc7061-cbbe-459e-bcde-09d1c700ceba","Type":"ContainerStarted","Data":"0babd88b77dfff447abc279c6c789f40820a50a78c4f6d6763a1f46f4f66bc97"} Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.068496 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-b9bnh" podStartSLOduration=123.068472257 podStartE2EDuration="2m3.068472257s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:46.949740902 +0000 UTC m=+143.284245569" watchObservedRunningTime="2025-12-09 09:49:47.068472257 +0000 UTC m=+143.402976924" Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.108325 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjsbv" event={"ID":"a4e4cf9c-48df-43b1-8ad7-9826c6220a45","Type":"ContainerStarted","Data":"8cceb930eda122c0ffb675305b2c7c987ec0a4bb5621e77935eb104e8e118f55"} Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.130365 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-847gg" podStartSLOduration=123.130347071 podStartE2EDuration="2m3.130347071s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:47.110603395 +0000 UTC m=+143.445108072" watchObservedRunningTime="2025-12-09 09:49:47.130347071 +0000 UTC m=+143.464851738" Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.132132 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jgwrj" podStartSLOduration=123.132125249 podStartE2EDuration="2m3.132125249s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:47.069736999 +0000 UTC m=+143.404241666" watchObservedRunningTime="2025-12-09 09:49:47.132125249 +0000 UTC m=+143.466629926" Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.135674 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:47 crc kubenswrapper[4824]: E1209 09:49:47.137727 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:47.637693862 +0000 UTC m=+143.972198709 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.165897 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjsbv" podStartSLOduration=123.165877704 podStartE2EDuration="2m3.165877704s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:47.1633236 +0000 UTC m=+143.497828267" watchObservedRunningTime="2025-12-09 09:49:47.165877704 +0000 UTC m=+143.500382371" Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.171431 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" event={"ID":"6c10edf9-cf75-4633-8d4c-9bbb9d9b6339","Type":"ContainerStarted","Data":"6a97ff98745208ecd29021b1cb9ab7211e4a073009dc1e1c4aa351efcb0a75a9"} Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.204238 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zjjp6" event={"ID":"2474e03a-283f-48af-ac3a-eebe08168b8f","Type":"ContainerStarted","Data":"5bcdfd038dfad62d26347b9b0f76373cb8910c5ec1ec959f6d8d646c8961dd46"} Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.234422 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" event={"ID":"7b26f121-04f6-4501-84b2-1833b927aa14","Type":"ContainerStarted","Data":"3efabae52732fd1dd695c8ec9ef93d6237ea498a410bac3039995980f09cd0d7"} Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.235358 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.251712 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:47 crc kubenswrapper[4824]: E1209 09:49:47.256112 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:47.756091076 +0000 UTC m=+144.090595743 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.285589 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-mj7jx" event={"ID":"dac1ac9b-4ddd-41b2-813b-ed7d8a71642f","Type":"ContainerStarted","Data":"b81f987a695aac3b20392984022ac9b8c31c087c6d4439490168cb2c88e811c7"} Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.298118 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" podStartSLOduration=123.2980959 podStartE2EDuration="2m3.2980959s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:47.297152959 +0000 UTC m=+143.631657636" watchObservedRunningTime="2025-12-09 09:49:47.2980959 +0000 UTC m=+143.632600567" Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.329253 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4rv7m" event={"ID":"14ebfe85-ea7c-49a4-a2e2-f8822c3a111e","Type":"ContainerStarted","Data":"eb0fd94dd5db7d7e9a05603efc0c6d75e04acc943fe8583bc0458ba7d05b9487"} Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.359053 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:47 crc kubenswrapper[4824]: E1209 09:49:47.360597 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:47.860577645 +0000 UTC m=+144.195082302 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.390303 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-4rv7m" podStartSLOduration=123.390274377 podStartE2EDuration="2m3.390274377s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:47.388431377 +0000 UTC m=+143.722936054" watchObservedRunningTime="2025-12-09 09:49:47.390274377 +0000 UTC m=+143.724779044" Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.415004 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-pk58m" event={"ID":"648e0288-ccba-422a-a8c8-b41d12a20845","Type":"ContainerStarted","Data":"327f85f3322991b583b640d4b320989209993b66b9c6ef565e35cd2084a894e7"} Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.463166 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:47 crc kubenswrapper[4824]: E1209 09:49:47.463677 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:47.963650948 +0000 UTC m=+144.298155625 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.565227 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:47 crc kubenswrapper[4824]: E1209 09:49:47.566963 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:48.066927317 +0000 UTC m=+144.401431984 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.649360 4824 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-tg5lv container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.649431 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" podUID="45732d49-ce14-439b-8b7a-fb9fdf267fb1" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.668004 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:47 crc kubenswrapper[4824]: E1209 09:49:47.673884 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:48.173860036 +0000 UTC m=+144.508364703 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.770445 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:47 crc kubenswrapper[4824]: E1209 09:49:47.771212 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:48.271189921 +0000 UTC m=+144.605694588 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.862976 4824 patch_prober.go:28] interesting pod/router-default-5444994796-mj7jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 09:49:47 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Dec 09 09:49:47 crc kubenswrapper[4824]: [+]process-running ok Dec 09 09:49:47 crc kubenswrapper[4824]: healthz check failed Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.863035 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.872435 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:47 crc kubenswrapper[4824]: E1209 09:49:47.872816 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:48.372799836 +0000 UTC m=+144.707304503 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:47 crc kubenswrapper[4824]: I1209 09:49:47.973416 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:47 crc kubenswrapper[4824]: E1209 09:49:47.973818 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:48.473798831 +0000 UTC m=+144.808303498 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.083474 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:48 crc kubenswrapper[4824]: E1209 09:49:48.084219 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:48.584205193 +0000 UTC m=+144.918709860 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.184519 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:48 crc kubenswrapper[4824]: E1209 09:49:48.185061 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:48.685037243 +0000 UTC m=+145.019541910 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.286489 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:48 crc kubenswrapper[4824]: E1209 09:49:48.287068 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:48.787048921 +0000 UTC m=+145.121553598 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.388935 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:48 crc kubenswrapper[4824]: E1209 09:49:48.389244 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:48.889203133 +0000 UTC m=+145.223707800 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.389704 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:48 crc kubenswrapper[4824]: E1209 09:49:48.390249 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:48.890229367 +0000 UTC m=+145.224734034 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.425369 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421225-96fgb" event={"ID":"055ead87-3e57-4053-a026-24566879437c","Type":"ContainerStarted","Data":"91eeeddf7380a1f76b21a29c04f33b21a23c21087506d6559bdb21b8ad896cac"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.425438 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421225-96fgb" event={"ID":"055ead87-3e57-4053-a026-24566879437c","Type":"ContainerStarted","Data":"636e96bbd2a46a602b57695d044dc991d80616bee0bacea2972ec7bba0e16bff"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.447683 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pvgjx" event={"ID":"ab5c6701-8122-4d9e-829d-93018eb67d7b","Type":"ContainerStarted","Data":"52a721d3fe9c19c6175db8f0a6b6443490dc067b258ab3a7cbf13deb01e87fea"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.447745 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pvgjx" event={"ID":"ab5c6701-8122-4d9e-829d-93018eb67d7b","Type":"ContainerStarted","Data":"ab2b914ccc43f09900852abed80cab531f64752f8e403ac494cf8be878ecc8f5"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.453337 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-mnksj" event={"ID":"67324e78-0fdb-4c67-a755-e2277801c106","Type":"ContainerStarted","Data":"80da93bbcff0a81376359343d8570cf4b3f4acd7530201e5aa2801be9237676c"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.453385 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-mnksj" event={"ID":"67324e78-0fdb-4c67-a755-e2277801c106","Type":"ContainerStarted","Data":"e60a722d6b9a869b60da4ceb986db19fb57f3accdd2825385d9f76455a3f9341"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.465713 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29421225-96fgb" podStartSLOduration=124.465689437 podStartE2EDuration="2m4.465689437s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:48.463291098 +0000 UTC m=+144.797795775" watchObservedRunningTime="2025-12-09 09:49:48.465689437 +0000 UTC m=+144.800194104" Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.466480 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-pk58m" podStartSLOduration=124.466473473 podStartE2EDuration="2m4.466473473s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:47.453146494 +0000 UTC m=+143.787651161" watchObservedRunningTime="2025-12-09 09:49:48.466473473 +0000 UTC m=+144.800978140" Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.497228 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:48 crc kubenswrapper[4824]: E1209 09:49:48.497506 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:48.997468017 +0000 UTC m=+145.331972684 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.497729 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:48 crc kubenswrapper[4824]: E1209 09:49:48.498190 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:48.99818324 +0000 UTC m=+145.332687907 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.499340 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-c9nzk" event={"ID":"401617ee-8306-4176-9af1-4ed2a5f9af0d","Type":"ContainerStarted","Data":"5051e078c58ef38eb914d0ea7f8d815f1ecb0eb579685c88c58a7f9e12817f2f"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.499387 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-c9nzk" event={"ID":"401617ee-8306-4176-9af1-4ed2a5f9af0d","Type":"ContainerStarted","Data":"8daec8a03cad694512821b457654416e4e50fcc49fe7561ac37ccd2fc6c0662d"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.505081 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-mnksj" podStartSLOduration=124.505053005 podStartE2EDuration="2m4.505053005s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:48.486202198 +0000 UTC m=+144.820706875" watchObservedRunningTime="2025-12-09 09:49:48.505053005 +0000 UTC m=+144.839557672" Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.558156 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zl4mr" event={"ID":"0f899ebd-d4e0-4e7c-82ea-914abad9dab6","Type":"ContainerStarted","Data":"3446fabf2fa7a3b06883197e53afc7fddecb1781a6e773e0928bc03c8072557a"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.558211 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zl4mr" event={"ID":"0f899ebd-d4e0-4e7c-82ea-914abad9dab6","Type":"ContainerStarted","Data":"8ac18bbfa3858ec2109312c88907bdef2b2633e812736ac123a759ff3126dcd9"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.558224 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zl4mr" event={"ID":"0f899ebd-d4e0-4e7c-82ea-914abad9dab6","Type":"ContainerStarted","Data":"9a9af78594ca40b933a3126698b02404fb7bafd5a5960aca8d88f1cabdfb13ed"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.574477 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pvgjx" podStartSLOduration=124.574455366 podStartE2EDuration="2m4.574455366s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:48.524318885 +0000 UTC m=+144.858823552" watchObservedRunningTime="2025-12-09 09:49:48.574455366 +0000 UTC m=+144.908960033" Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.576104 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-c9nzk" podStartSLOduration=9.57609627 podStartE2EDuration="9.57609627s" podCreationTimestamp="2025-12-09 09:49:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:48.572562404 +0000 UTC m=+144.907067071" watchObservedRunningTime="2025-12-09 09:49:48.57609627 +0000 UTC m=+144.910600937" Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.577155 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-pk58m" event={"ID":"648e0288-ccba-422a-a8c8-b41d12a20845","Type":"ContainerStarted","Data":"2c730a4af34f538ce9019329bcf989a2cc0db283bb04ea357df396cc031fbf5a"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.585158 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" event={"ID":"6c10edf9-cf75-4633-8d4c-9bbb9d9b6339","Type":"ContainerStarted","Data":"886bc9d74183124b07bb450444597fe02f70436ac80f07efff6ededc8ff84914"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.585222 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" event={"ID":"6c10edf9-cf75-4633-8d4c-9bbb9d9b6339","Type":"ContainerStarted","Data":"e43d3133a420e9d81727e2dc483c97831422fbc73da9250b182f12fc2de78127"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.587398 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.594483 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zl4mr" podStartSLOduration=124.594462961 podStartE2EDuration="2m4.594462961s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:48.59108378 +0000 UTC m=+144.925588447" watchObservedRunningTime="2025-12-09 09:49:48.594462961 +0000 UTC m=+144.928967628" Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.598628 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:48 crc kubenswrapper[4824]: E1209 09:49:48.598878 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:49.098857164 +0000 UTC m=+145.433361831 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.625106 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" podStartSLOduration=124.625090693 podStartE2EDuration="2m4.625090693s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:48.624239124 +0000 UTC m=+144.958743791" watchObservedRunningTime="2025-12-09 09:49:48.625090693 +0000 UTC m=+144.959595360" Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.633674 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zjjp6" event={"ID":"2474e03a-283f-48af-ac3a-eebe08168b8f","Type":"ContainerStarted","Data":"a8cb59a6fd213e314fdf296fe75f4e77ff77318c27db952dbd6747167b4a4039"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.633939 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zjjp6" event={"ID":"2474e03a-283f-48af-ac3a-eebe08168b8f","Type":"ContainerStarted","Data":"2cdf373c53a5781d9be616ec139fa6801b8fb25c00799370d4c44f0d587adb0d"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.634793 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-zjjp6" Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.653869 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" event={"ID":"c0ac631f-5d13-4814-84d3-61f7bd704f9e","Type":"ContainerStarted","Data":"d617a8f77bcf05f918bdee243c82993abbe105e5f896eac676b779b5e6cce789"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.675009 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kv8sn" event={"ID":"54818ac0-7b5e-452f-b2f7-6a87a667a76f","Type":"ContainerStarted","Data":"05c140427626f12ded7705e709b041e27d2bc8ca00d4082484def3750ef4f06b"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.675089 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kv8sn" event={"ID":"54818ac0-7b5e-452f-b2f7-6a87a667a76f","Type":"ContainerStarted","Data":"9c2c0529106a07b565471232591095d533d8728797d7f5fe4875267115162a8b"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.676743 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-kv8sn" Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.689422 4824 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-kv8sn container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.689743 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-kv8sn" podUID="54818ac0-7b5e-452f-b2f7-6a87a667a76f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.692910 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-b9bnh" event={"ID":"6efb8245-07e9-4c31-b8ab-53ab3685e593","Type":"ContainerStarted","Data":"40e458dc5060ba615b1ead2dedddbf92898e7b6955c26c74843145945afc3ed4"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.694343 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-b9bnh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.694405 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b9bnh" podUID="6efb8245-07e9-4c31-b8ab-53ab3685e593" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.700606 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:48 crc kubenswrapper[4824]: E1209 09:49:48.702448 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:49.202429833 +0000 UTC m=+145.536934500 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.708584 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmk4v" event={"ID":"564831de-2ff3-4051-a1a0-c5317bf10ca1","Type":"ContainerStarted","Data":"9736f6248104e9d8e34f196c30b58db107259f49947e646d6cea79fd8d139ed5"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.720379 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.732844 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-zjjp6" podStartSLOduration=9.732823968 podStartE2EDuration="9.732823968s" podCreationTimestamp="2025-12-09 09:49:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:48.671058957 +0000 UTC m=+145.005563624" watchObservedRunningTime="2025-12-09 09:49:48.732823968 +0000 UTC m=+145.067328635" Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.733291 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-kv8sn" podStartSLOduration=124.733285933 podStartE2EDuration="2m4.733285933s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:48.732075764 +0000 UTC m=+145.066580441" watchObservedRunningTime="2025-12-09 09:49:48.733285933 +0000 UTC m=+145.067790600" Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.739361 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hjg74" event={"ID":"37c820ff-b4bf-41ae-bcf3-8e0f7368bb12","Type":"ContainerStarted","Data":"261f10de356de8e35c324f44663526e5b3c78fca5123ac95e754109a41510071"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.739419 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hjg74" event={"ID":"37c820ff-b4bf-41ae-bcf3-8e0f7368bb12","Type":"ContainerStarted","Data":"136f9e4e0ae231b7bf612195ad19ad05606991bfa3887780eb960b9d26e3cc70"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.739431 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hjg74" event={"ID":"37c820ff-b4bf-41ae-bcf3-8e0f7368bb12","Type":"ContainerStarted","Data":"f2584400dffcf142ec370abed9269795f98e222d57ee1a31a4164ae31b045f52"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.752317 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-99q6g" event={"ID":"7a94cb6d-599c-4dae-b7db-43f67a6620b6","Type":"ContainerStarted","Data":"806552b0c01c3c01549f8c84dbe5df2917ba7830c1be8e7cee386ed72a6c5165"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.752364 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-99q6g" event={"ID":"7a94cb6d-599c-4dae-b7db-43f67a6620b6","Type":"ContainerStarted","Data":"88f77d1cf4ee34edc771c7caa00ae01fc04d7dbf2e6258d8611de3de57fa0703"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.768701 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q8c9m" event={"ID":"eb2d4bd5-8dac-4b9d-9125-51fca045c2b3","Type":"ContainerStarted","Data":"c71c48cfd13208fb9ad3c5859792bc57503ebef3f3075afb8ac88446eb844499"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.768767 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q8c9m" event={"ID":"eb2d4bd5-8dac-4b9d-9125-51fca045c2b3","Type":"ContainerStarted","Data":"01e3952140a60d1e9bada3d1f435f0ecb58cdfd801e29e2f6944b287f76383c6"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.778889 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmk4v" podStartSLOduration=124.778846384 podStartE2EDuration="2m4.778846384s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:48.772983302 +0000 UTC m=+145.107487989" watchObservedRunningTime="2025-12-09 09:49:48.778846384 +0000 UTC m=+145.113351061" Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.794508 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c2dfm" event={"ID":"a7c02c2b-70ae-4b6f-85ae-5a38b737a73b","Type":"ContainerStarted","Data":"c840b64467f77a57764598d2da12e8550523b9e8eb8e2f82efdbc54d84113ea5"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.794565 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c2dfm" event={"ID":"a7c02c2b-70ae-4b6f-85ae-5a38b737a73b","Type":"ContainerStarted","Data":"6d2f01748dc810b8b818832c4f3a8dc88c578493ff9ba39706d9b73f86486d8d"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.799935 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c2dfm" Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.811183 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:48 crc kubenswrapper[4824]: E1209 09:49:48.813471 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:49.313444526 +0000 UTC m=+145.647949193 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.827678 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pd82d" event={"ID":"7bff53ba-5406-445c-afcb-3247516a1258","Type":"ContainerStarted","Data":"b4e4cb3e6ff2d91fe99794540d1fc49cd99c764b110fc0b0c139a7a6581c816f"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.842161 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c2dfm" Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.870863 4824 patch_prober.go:28] interesting pod/router-default-5444994796-mj7jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 09:49:48 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Dec 09 09:49:48 crc kubenswrapper[4824]: [+]process-running ok Dec 09 09:49:48 crc kubenswrapper[4824]: healthz check failed Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.870980 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.891213 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" event={"ID":"62823972-71dc-4f7d-b4f6-da0cb90dcf36","Type":"ContainerStarted","Data":"e048faa53b6a634fed8d590a6500c0a2619064882a9070e599eea82d1a432a34"} Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.891289 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.914430 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:48 crc kubenswrapper[4824]: E1209 09:49:48.915792 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:49.415758354 +0000 UTC m=+145.750263021 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.952879 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.961958 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q8c9m" podStartSLOduration=124.961930795 podStartE2EDuration="2m4.961930795s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:48.954266634 +0000 UTC m=+145.288771311" watchObservedRunningTime="2025-12-09 09:49:48.961930795 +0000 UTC m=+145.296435462" Dec 09 09:49:48 crc kubenswrapper[4824]: I1209 09:49:48.963425 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-99q6g" podStartSLOduration=124.963414503 podStartE2EDuration="2m4.963414503s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:48.913341105 +0000 UTC m=+145.247845772" watchObservedRunningTime="2025-12-09 09:49:48.963414503 +0000 UTC m=+145.297919170" Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.003939 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.010813 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hjg74" podStartSLOduration=125.010753912 podStartE2EDuration="2m5.010753912s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:48.983991857 +0000 UTC m=+145.318496534" watchObservedRunningTime="2025-12-09 09:49:49.010753912 +0000 UTC m=+145.345258579" Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.017372 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:49 crc kubenswrapper[4824]: E1209 09:49:49.017656 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:49.517615147 +0000 UTC m=+145.852119884 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.018374 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:49 crc kubenswrapper[4824]: E1209 09:49:49.019717 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:49.519695955 +0000 UTC m=+145.854200842 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.025284 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c2dfm" podStartSLOduration=125.025268057 podStartE2EDuration="2m5.025268057s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:49.024717629 +0000 UTC m=+145.359222296" watchObservedRunningTime="2025-12-09 09:49:49.025268057 +0000 UTC m=+145.359772714" Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.119713 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:49 crc kubenswrapper[4824]: E1209 09:49:49.120330 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:49.620310327 +0000 UTC m=+145.954814994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.139174 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" podStartSLOduration=125.139154584 podStartE2EDuration="2m5.139154584s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:49.135312718 +0000 UTC m=+145.469817405" watchObservedRunningTime="2025-12-09 09:49:49.139154584 +0000 UTC m=+145.473659241" Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.221725 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:49 crc kubenswrapper[4824]: E1209 09:49:49.222155 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:49.72214204 +0000 UTC m=+146.056646707 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.323202 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:49 crc kubenswrapper[4824]: E1209 09:49:49.323353 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:49.823327301 +0000 UTC m=+146.157831968 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.323533 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:49 crc kubenswrapper[4824]: E1209 09:49:49.323977 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:49.823967171 +0000 UTC m=+146.158471838 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.424328 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:49 crc kubenswrapper[4824]: E1209 09:49:49.424482 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:49.92445677 +0000 UTC m=+146.258961437 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.425029 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:49 crc kubenswrapper[4824]: E1209 09:49:49.425319 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:49.925312198 +0000 UTC m=+146.259816865 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.525756 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:49 crc kubenswrapper[4824]: E1209 09:49:49.526059 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:50.025999202 +0000 UTC m=+146.360503879 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.526545 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:49 crc kubenswrapper[4824]: E1209 09:49:49.526950 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:50.026927463 +0000 UTC m=+146.361432230 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.596849 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9pdnx"] Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.598117 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9pdnx" Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.600947 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.623249 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9pdnx"] Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.627936 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:49 crc kubenswrapper[4824]: E1209 09:49:49.628214 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:50.128180056 +0000 UTC m=+146.462684723 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.628592 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:49 crc kubenswrapper[4824]: E1209 09:49:49.628992 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:50.128973101 +0000 UTC m=+146.463477768 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.729306 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.729582 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbkw6\" (UniqueName: \"kubernetes.io/projected/db041858-ed93-4df9-a5d2-f79b7f9c3275-kube-api-access-vbkw6\") pod \"certified-operators-9pdnx\" (UID: \"db041858-ed93-4df9-a5d2-f79b7f9c3275\") " pod="openshift-marketplace/certified-operators-9pdnx" Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.729622 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db041858-ed93-4df9-a5d2-f79b7f9c3275-utilities\") pod \"certified-operators-9pdnx\" (UID: \"db041858-ed93-4df9-a5d2-f79b7f9c3275\") " pod="openshift-marketplace/certified-operators-9pdnx" Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.729690 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db041858-ed93-4df9-a5d2-f79b7f9c3275-catalog-content\") pod \"certified-operators-9pdnx\" (UID: \"db041858-ed93-4df9-a5d2-f79b7f9c3275\") " pod="openshift-marketplace/certified-operators-9pdnx" Dec 09 09:49:49 crc kubenswrapper[4824]: E1209 09:49:49.729827 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:50.229801211 +0000 UTC m=+146.564305878 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.800319 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zmxrv"] Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.801377 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zmxrv" Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.805083 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.820405 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zmxrv"] Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.830814 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db041858-ed93-4df9-a5d2-f79b7f9c3275-catalog-content\") pod \"certified-operators-9pdnx\" (UID: \"db041858-ed93-4df9-a5d2-f79b7f9c3275\") " pod="openshift-marketplace/certified-operators-9pdnx" Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.830940 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbkw6\" (UniqueName: \"kubernetes.io/projected/db041858-ed93-4df9-a5d2-f79b7f9c3275-kube-api-access-vbkw6\") pod \"certified-operators-9pdnx\" (UID: \"db041858-ed93-4df9-a5d2-f79b7f9c3275\") " pod="openshift-marketplace/certified-operators-9pdnx" Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.830979 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db041858-ed93-4df9-a5d2-f79b7f9c3275-utilities\") pod \"certified-operators-9pdnx\" (UID: \"db041858-ed93-4df9-a5d2-f79b7f9c3275\") " pod="openshift-marketplace/certified-operators-9pdnx" Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.831030 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:49 crc kubenswrapper[4824]: E1209 09:49:49.831448 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:50.331428707 +0000 UTC m=+146.665933374 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.831474 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db041858-ed93-4df9-a5d2-f79b7f9c3275-catalog-content\") pod \"certified-operators-9pdnx\" (UID: \"db041858-ed93-4df9-a5d2-f79b7f9c3275\") " pod="openshift-marketplace/certified-operators-9pdnx" Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.831860 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db041858-ed93-4df9-a5d2-f79b7f9c3275-utilities\") pod \"certified-operators-9pdnx\" (UID: \"db041858-ed93-4df9-a5d2-f79b7f9c3275\") " pod="openshift-marketplace/certified-operators-9pdnx" Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.841421 4824 patch_prober.go:28] interesting pod/router-default-5444994796-mj7jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 09:49:49 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Dec 09 09:49:49 crc kubenswrapper[4824]: [+]process-running ok Dec 09 09:49:49 crc kubenswrapper[4824]: healthz check failed Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.841482 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.866211 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbkw6\" (UniqueName: \"kubernetes.io/projected/db041858-ed93-4df9-a5d2-f79b7f9c3275-kube-api-access-vbkw6\") pod \"certified-operators-9pdnx\" (UID: \"db041858-ed93-4df9-a5d2-f79b7f9c3275\") " pod="openshift-marketplace/certified-operators-9pdnx" Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.913636 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9pdnx" Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.921884 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pd82d" event={"ID":"7bff53ba-5406-445c-afcb-3247516a1258","Type":"ContainerStarted","Data":"12d3f0edc45a316a3ee8570b7b399d21a935cb5d32f4b23ad7e5654f39c734c7"} Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.933290 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.933613 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28b6h\" (UniqueName: \"kubernetes.io/projected/e799b94c-33db-4de6-afd4-a8ffc4e9b069-kube-api-access-28b6h\") pod \"community-operators-zmxrv\" (UID: \"e799b94c-33db-4de6-afd4-a8ffc4e9b069\") " pod="openshift-marketplace/community-operators-zmxrv" Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.933672 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e799b94c-33db-4de6-afd4-a8ffc4e9b069-utilities\") pod \"community-operators-zmxrv\" (UID: \"e799b94c-33db-4de6-afd4-a8ffc4e9b069\") " pod="openshift-marketplace/community-operators-zmxrv" Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.933729 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e799b94c-33db-4de6-afd4-a8ffc4e9b069-catalog-content\") pod \"community-operators-zmxrv\" (UID: \"e799b94c-33db-4de6-afd4-a8ffc4e9b069\") " pod="openshift-marketplace/community-operators-zmxrv" Dec 09 09:49:49 crc kubenswrapper[4824]: E1209 09:49:49.933871 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:50.433845458 +0000 UTC m=+146.768350125 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.951121 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-mnksj" event={"ID":"67324e78-0fdb-4c67-a755-e2277801c106","Type":"ContainerStarted","Data":"6c34cde772ecd9026d1f908330d25fef1c5ff839ab7c95ebdd942dbd51f4faf5"} Dec 09 09:49:49 crc kubenswrapper[4824]: I1209 09:49:49.986503 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-pd82d" podStartSLOduration=125.98647531 podStartE2EDuration="2m5.98647531s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:49.983376689 +0000 UTC m=+146.317881356" watchObservedRunningTime="2025-12-09 09:49:49.98647531 +0000 UTC m=+146.320979977" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.005202 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" event={"ID":"c0ac631f-5d13-4814-84d3-61f7bd704f9e","Type":"ContainerStarted","Data":"9e133cb75e3e1728a5a1f63d956a8d5739c8ed28e6b2f1fb8eb7ad1b4e6d114b"} Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.005284 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" event={"ID":"c0ac631f-5d13-4814-84d3-61f7bd704f9e","Type":"ContainerStarted","Data":"1d1f74af8efa951be937df3c20063021235296c303f3cfc3729b0e0a63db8aac"} Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.007519 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-g9g6w"] Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.008837 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g9g6w" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.016100 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmk4v" event={"ID":"564831de-2ff3-4051-a1a0-c5317bf10ca1","Type":"ContainerStarted","Data":"c59080717ba80630f00197423c33b8d879b8ee6ad2220e75a1e42b0e24aedbc9"} Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.023023 4824 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-kv8sn container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.023066 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-b9bnh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.023121 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-kv8sn" podUID="54818ac0-7b5e-452f-b2f7-6a87a667a76f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.023133 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b9bnh" podUID="6efb8245-07e9-4c31-b8ab-53ab3685e593" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.072092 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.072192 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e799b94c-33db-4de6-afd4-a8ffc4e9b069-catalog-content\") pod \"community-operators-zmxrv\" (UID: \"e799b94c-33db-4de6-afd4-a8ffc4e9b069\") " pod="openshift-marketplace/community-operators-zmxrv" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.072459 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28b6h\" (UniqueName: \"kubernetes.io/projected/e799b94c-33db-4de6-afd4-a8ffc4e9b069-kube-api-access-28b6h\") pod \"community-operators-zmxrv\" (UID: \"e799b94c-33db-4de6-afd4-a8ffc4e9b069\") " pod="openshift-marketplace/community-operators-zmxrv" Dec 09 09:49:50 crc kubenswrapper[4824]: E1209 09:49:50.082504 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:50.582482131 +0000 UTC m=+146.916986798 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.083051 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e799b94c-33db-4de6-afd4-a8ffc4e9b069-catalog-content\") pod \"community-operators-zmxrv\" (UID: \"e799b94c-33db-4de6-afd4-a8ffc4e9b069\") " pod="openshift-marketplace/community-operators-zmxrv" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.094653 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e799b94c-33db-4de6-afd4-a8ffc4e9b069-utilities\") pod \"community-operators-zmxrv\" (UID: \"e799b94c-33db-4de6-afd4-a8ffc4e9b069\") " pod="openshift-marketplace/community-operators-zmxrv" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.096337 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e799b94c-33db-4de6-afd4-a8ffc4e9b069-utilities\") pod \"community-operators-zmxrv\" (UID: \"e799b94c-33db-4de6-afd4-a8ffc4e9b069\") " pod="openshift-marketplace/community-operators-zmxrv" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.108418 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g9g6w"] Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.150673 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28b6h\" (UniqueName: \"kubernetes.io/projected/e799b94c-33db-4de6-afd4-a8ffc4e9b069-kube-api-access-28b6h\") pod \"community-operators-zmxrv\" (UID: \"e799b94c-33db-4de6-afd4-a8ffc4e9b069\") " pod="openshift-marketplace/community-operators-zmxrv" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.203716 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.204294 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/741767ab-ed9f-4b68-83bb-41aa25497741-utilities\") pod \"certified-operators-g9g6w\" (UID: \"741767ab-ed9f-4b68-83bb-41aa25497741\") " pod="openshift-marketplace/certified-operators-g9g6w" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.204367 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/741767ab-ed9f-4b68-83bb-41aa25497741-catalog-content\") pod \"certified-operators-g9g6w\" (UID: \"741767ab-ed9f-4b68-83bb-41aa25497741\") " pod="openshift-marketplace/certified-operators-g9g6w" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.204658 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wswd5\" (UniqueName: \"kubernetes.io/projected/741767ab-ed9f-4b68-83bb-41aa25497741-kube-api-access-wswd5\") pod \"certified-operators-g9g6w\" (UID: \"741767ab-ed9f-4b68-83bb-41aa25497741\") " pod="openshift-marketplace/certified-operators-g9g6w" Dec 09 09:49:50 crc kubenswrapper[4824]: E1209 09:49:50.205737 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:50.705707173 +0000 UTC m=+147.040211840 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.210360 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xg5qv"] Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.221350 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xg5qv" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.246233 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xg5qv"] Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.325072 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.325137 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad53b3fe-0e9e-4d56-9ca0-212dde3ac669-utilities\") pod \"community-operators-xg5qv\" (UID: \"ad53b3fe-0e9e-4d56-9ca0-212dde3ac669\") " pod="openshift-marketplace/community-operators-xg5qv" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.325166 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/741767ab-ed9f-4b68-83bb-41aa25497741-utilities\") pod \"certified-operators-g9g6w\" (UID: \"741767ab-ed9f-4b68-83bb-41aa25497741\") " pod="openshift-marketplace/certified-operators-g9g6w" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.325194 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/741767ab-ed9f-4b68-83bb-41aa25497741-catalog-content\") pod \"certified-operators-g9g6w\" (UID: \"741767ab-ed9f-4b68-83bb-41aa25497741\") " pod="openshift-marketplace/certified-operators-g9g6w" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.325253 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad53b3fe-0e9e-4d56-9ca0-212dde3ac669-catalog-content\") pod \"community-operators-xg5qv\" (UID: \"ad53b3fe-0e9e-4d56-9ca0-212dde3ac669\") " pod="openshift-marketplace/community-operators-xg5qv" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.325284 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wswd5\" (UniqueName: \"kubernetes.io/projected/741767ab-ed9f-4b68-83bb-41aa25497741-kube-api-access-wswd5\") pod \"certified-operators-g9g6w\" (UID: \"741767ab-ed9f-4b68-83bb-41aa25497741\") " pod="openshift-marketplace/certified-operators-g9g6w" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.325302 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx6hv\" (UniqueName: \"kubernetes.io/projected/ad53b3fe-0e9e-4d56-9ca0-212dde3ac669-kube-api-access-qx6hv\") pod \"community-operators-xg5qv\" (UID: \"ad53b3fe-0e9e-4d56-9ca0-212dde3ac669\") " pod="openshift-marketplace/community-operators-xg5qv" Dec 09 09:49:50 crc kubenswrapper[4824]: E1209 09:49:50.325731 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:50.825716241 +0000 UTC m=+147.160220908 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.327200 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/741767ab-ed9f-4b68-83bb-41aa25497741-utilities\") pod \"certified-operators-g9g6w\" (UID: \"741767ab-ed9f-4b68-83bb-41aa25497741\") " pod="openshift-marketplace/certified-operators-g9g6w" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.327412 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/741767ab-ed9f-4b68-83bb-41aa25497741-catalog-content\") pod \"certified-operators-g9g6w\" (UID: \"741767ab-ed9f-4b68-83bb-41aa25497741\") " pod="openshift-marketplace/certified-operators-g9g6w" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.383544 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wswd5\" (UniqueName: \"kubernetes.io/projected/741767ab-ed9f-4b68-83bb-41aa25497741-kube-api-access-wswd5\") pod \"certified-operators-g9g6w\" (UID: \"741767ab-ed9f-4b68-83bb-41aa25497741\") " pod="openshift-marketplace/certified-operators-g9g6w" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.384029 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g9g6w" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.437498 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.438138 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad53b3fe-0e9e-4d56-9ca0-212dde3ac669-catalog-content\") pod \"community-operators-xg5qv\" (UID: \"ad53b3fe-0e9e-4d56-9ca0-212dde3ac669\") " pod="openshift-marketplace/community-operators-xg5qv" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.438198 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx6hv\" (UniqueName: \"kubernetes.io/projected/ad53b3fe-0e9e-4d56-9ca0-212dde3ac669-kube-api-access-qx6hv\") pod \"community-operators-xg5qv\" (UID: \"ad53b3fe-0e9e-4d56-9ca0-212dde3ac669\") " pod="openshift-marketplace/community-operators-xg5qv" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.438266 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad53b3fe-0e9e-4d56-9ca0-212dde3ac669-utilities\") pod \"community-operators-xg5qv\" (UID: \"ad53b3fe-0e9e-4d56-9ca0-212dde3ac669\") " pod="openshift-marketplace/community-operators-xg5qv" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.438406 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zmxrv" Dec 09 09:49:50 crc kubenswrapper[4824]: E1209 09:49:50.438869 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:50.938848803 +0000 UTC m=+147.273353460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.439540 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad53b3fe-0e9e-4d56-9ca0-212dde3ac669-utilities\") pod \"community-operators-xg5qv\" (UID: \"ad53b3fe-0e9e-4d56-9ca0-212dde3ac669\") " pod="openshift-marketplace/community-operators-xg5qv" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.439550 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad53b3fe-0e9e-4d56-9ca0-212dde3ac669-catalog-content\") pod \"community-operators-xg5qv\" (UID: \"ad53b3fe-0e9e-4d56-9ca0-212dde3ac669\") " pod="openshift-marketplace/community-operators-xg5qv" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.467657 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx6hv\" (UniqueName: \"kubernetes.io/projected/ad53b3fe-0e9e-4d56-9ca0-212dde3ac669-kube-api-access-qx6hv\") pod \"community-operators-xg5qv\" (UID: \"ad53b3fe-0e9e-4d56-9ca0-212dde3ac669\") " pod="openshift-marketplace/community-operators-xg5qv" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.517480 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9pdnx"] Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.539205 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:50 crc kubenswrapper[4824]: E1209 09:49:50.539574 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:51.039560028 +0000 UTC m=+147.374064685 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.544624 4824 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.613870 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xg5qv" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.641211 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:50 crc kubenswrapper[4824]: E1209 09:49:50.641548 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:51.141510615 +0000 UTC m=+147.476015282 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.641972 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:50 crc kubenswrapper[4824]: E1209 09:49:50.642392 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:51.142381493 +0000 UTC m=+147.476886160 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.742934 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:50 crc kubenswrapper[4824]: E1209 09:49:50.743729 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:51.243707098 +0000 UTC m=+147.578211765 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.812727 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zmxrv"] Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.839726 4824 patch_prober.go:28] interesting pod/router-default-5444994796-mj7jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 09:49:50 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Dec 09 09:49:50 crc kubenswrapper[4824]: [+]process-running ok Dec 09 09:49:50 crc kubenswrapper[4824]: healthz check failed Dec 09 09:49:50 crc kubenswrapper[4824]: W1209 09:49:50.839829 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode799b94c_33db_4de6_afd4_a8ffc4e9b069.slice/crio-0b490d86ca847167731a7a3b93a1f9cdbdc5cda854d96971826a82bd616519b0 WatchSource:0}: Error finding container 0b490d86ca847167731a7a3b93a1f9cdbdc5cda854d96971826a82bd616519b0: Status 404 returned error can't find the container with id 0b490d86ca847167731a7a3b93a1f9cdbdc5cda854d96971826a82bd616519b0 Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.839833 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.845435 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:50 crc kubenswrapper[4824]: E1209 09:49:50.845865 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:51.345844801 +0000 UTC m=+147.680349658 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.911897 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g9g6w"] Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.946587 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:50 crc kubenswrapper[4824]: E1209 09:49:50.946870 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:51.446833595 +0000 UTC m=+147.781338262 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.946938 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.947115 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.947169 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.947222 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.947403 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.950419 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:49:50 crc kubenswrapper[4824]: E1209 09:49:50.950794 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:51.450760683 +0000 UTC m=+147.785265540 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.954618 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.956261 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.959259 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:49:50 crc kubenswrapper[4824]: I1209 09:49:50.975646 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xg5qv"] Dec 09 09:49:50 crc kubenswrapper[4824]: W1209 09:49:50.976536 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad53b3fe_0e9e_4d56_9ca0_212dde3ac669.slice/crio-0b2834a7c7988d5f82185d739f6af31e95a8864c29673c0fba63dee5b8ffab79 WatchSource:0}: Error finding container 0b2834a7c7988d5f82185d739f6af31e95a8864c29673c0fba63dee5b8ffab79: Status 404 returned error can't find the container with id 0b2834a7c7988d5f82185d739f6af31e95a8864c29673c0fba63dee5b8ffab79 Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.030291 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" event={"ID":"c0ac631f-5d13-4814-84d3-61f7bd704f9e","Type":"ContainerStarted","Data":"4b3d41ff9160e56da01b60aa9bce9a86f0de81b635d8eba1b6129d350de1506c"} Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.030366 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" event={"ID":"c0ac631f-5d13-4814-84d3-61f7bd704f9e","Type":"ContainerStarted","Data":"7e96977bd30cae68e2dccf76a842ac7351bc0f9f10ee52bd16077b2eff9697f6"} Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.043343 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xg5qv" event={"ID":"ad53b3fe-0e9e-4d56-9ca0-212dde3ac669","Type":"ContainerStarted","Data":"0b2834a7c7988d5f82185d739f6af31e95a8864c29673c0fba63dee5b8ffab79"} Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.049532 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:51 crc kubenswrapper[4824]: E1209 09:49:51.050009 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:51.54997608 +0000 UTC m=+147.884480747 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.054440 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zmxrv" event={"ID":"e799b94c-33db-4de6-afd4-a8ffc4e9b069","Type":"ContainerStarted","Data":"0b490d86ca847167731a7a3b93a1f9cdbdc5cda854d96971826a82bd616519b0"} Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.055650 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g9g6w" event={"ID":"741767ab-ed9f-4b68-83bb-41aa25497741","Type":"ContainerStarted","Data":"f33e5882217416bbc7a93a824c8b1bb99d2c6a6b81b5027edaa71d9882bccf22"} Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.057197 4824 generic.go:334] "Generic (PLEG): container finished" podID="db041858-ed93-4df9-a5d2-f79b7f9c3275" containerID="2962bf05684f13cc5e6a973a03e16b20b1b2afcf9b7dba35c04ac979fdf1a4bb" exitCode=0 Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.058268 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9pdnx" event={"ID":"db041858-ed93-4df9-a5d2-f79b7f9c3275","Type":"ContainerDied","Data":"2962bf05684f13cc5e6a973a03e16b20b1b2afcf9b7dba35c04ac979fdf1a4bb"} Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.058294 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9pdnx" event={"ID":"db041858-ed93-4df9-a5d2-f79b7f9c3275","Type":"ContainerStarted","Data":"b8ce610308575d8c5358c5bea7b4c7667f271d057920f405b1d3c55dcf210509"} Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.061773 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.070968 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-kv8sn" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.076270 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" podStartSLOduration=12.07623734 podStartE2EDuration="12.07623734s" podCreationTimestamp="2025-12-09 09:49:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:51.072824348 +0000 UTC m=+147.407329015" watchObservedRunningTime="2025-12-09 09:49:51.07623734 +0000 UTC m=+147.410742007" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.124958 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.134434 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.138948 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.150848 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:51 crc kubenswrapper[4824]: E1209 09:49:51.152890 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 09:49:51.652867927 +0000 UTC m=+147.987372654 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkdx" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.253893 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:51 crc kubenswrapper[4824]: E1209 09:49:51.254366 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 09:49:51.754341678 +0000 UTC m=+148.088846345 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.287124 4824 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-09T09:49:50.544642724Z","Handler":null,"Name":""} Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.315705 4824 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.315746 4824 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.355832 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.358429 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.358459 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.400756 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkdx\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.456991 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.472816 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.475146 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.478375 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.478532 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.491573 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:51 crc kubenswrapper[4824]: W1209 09:49:51.509768 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-76e3fe0dcc1522e50422a7b99213f8628c34794e39f950cabc892e5e4cf8d8e0 WatchSource:0}: Error finding container 76e3fe0dcc1522e50422a7b99213f8628c34794e39f950cabc892e5e4cf8d8e0: Status 404 returned error can't find the container with id 76e3fe0dcc1522e50422a7b99213f8628c34794e39f950cabc892e5e4cf8d8e0 Dec 09 09:49:51 crc kubenswrapper[4824]: W1209 09:49:51.532636 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-3230d69ff6f0b93927baf124b89a39b9f013e76ba9f10b8451852c81a15698a3 WatchSource:0}: Error finding container 3230d69ff6f0b93927baf124b89a39b9f013e76ba9f10b8451852c81a15698a3: Status 404 returned error can't find the container with id 3230d69ff6f0b93927baf124b89a39b9f013e76ba9f10b8451852c81a15698a3 Dec 09 09:49:51 crc kubenswrapper[4824]: W1209 09:49:51.553275 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-b73c6131f2328df15573a55c4aed52935b729e0d66a7351ae1fab6f69829eb90 WatchSource:0}: Error finding container b73c6131f2328df15573a55c4aed52935b729e0d66a7351ae1fab6f69829eb90: Status 404 returned error can't find the container with id b73c6131f2328df15573a55c4aed52935b729e0d66a7351ae1fab6f69829eb90 Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.743560 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qkkdx"] Dec 09 09:49:51 crc kubenswrapper[4824]: W1209 09:49:51.756773 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d910d95_4961_4d28_9599_087153ce7d2f.slice/crio-1b5175598a7d3f9dc435aec37623ff6fed9118de27a47b9b4509953362a3348a WatchSource:0}: Error finding container 1b5175598a7d3f9dc435aec37623ff6fed9118de27a47b9b4509953362a3348a: Status 404 returned error can't find the container with id 1b5175598a7d3f9dc435aec37623ff6fed9118de27a47b9b4509953362a3348a Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.786325 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-j7pg7"] Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.787349 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j7pg7" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.789524 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.815490 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j7pg7"] Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.838124 4824 patch_prober.go:28] interesting pod/router-default-5444994796-mj7jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 09:49:51 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Dec 09 09:49:51 crc kubenswrapper[4824]: [+]process-running ok Dec 09 09:49:51 crc kubenswrapper[4824]: healthz check failed Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.838195 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.862690 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58792ed4-e360-4827-bf8c-508146205b78-catalog-content\") pod \"redhat-marketplace-j7pg7\" (UID: \"58792ed4-e360-4827-bf8c-508146205b78\") " pod="openshift-marketplace/redhat-marketplace-j7pg7" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.862904 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58792ed4-e360-4827-bf8c-508146205b78-utilities\") pod \"redhat-marketplace-j7pg7\" (UID: \"58792ed4-e360-4827-bf8c-508146205b78\") " pod="openshift-marketplace/redhat-marketplace-j7pg7" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.862976 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jz54p\" (UniqueName: \"kubernetes.io/projected/58792ed4-e360-4827-bf8c-508146205b78-kube-api-access-jz54p\") pod \"redhat-marketplace-j7pg7\" (UID: \"58792ed4-e360-4827-bf8c-508146205b78\") " pod="openshift-marketplace/redhat-marketplace-j7pg7" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.905638 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.906013 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.908209 4824 patch_prober.go:28] interesting pod/console-f9d7485db-ldw7r container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.30:8443/health\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.908259 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-ldw7r" podUID="0c5e7133-21dc-46e0-b334-9d36c0427aa3" containerName="console" probeResult="failure" output="Get \"https://10.217.0.30:8443/health\": dial tcp 10.217.0.30:8443: connect: connection refused" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.918536 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.963666 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jz54p\" (UniqueName: \"kubernetes.io/projected/58792ed4-e360-4827-bf8c-508146205b78-kube-api-access-jz54p\") pod \"redhat-marketplace-j7pg7\" (UID: \"58792ed4-e360-4827-bf8c-508146205b78\") " pod="openshift-marketplace/redhat-marketplace-j7pg7" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.963721 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58792ed4-e360-4827-bf8c-508146205b78-catalog-content\") pod \"redhat-marketplace-j7pg7\" (UID: \"58792ed4-e360-4827-bf8c-508146205b78\") " pod="openshift-marketplace/redhat-marketplace-j7pg7" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.963840 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58792ed4-e360-4827-bf8c-508146205b78-utilities\") pod \"redhat-marketplace-j7pg7\" (UID: \"58792ed4-e360-4827-bf8c-508146205b78\") " pod="openshift-marketplace/redhat-marketplace-j7pg7" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.964738 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58792ed4-e360-4827-bf8c-508146205b78-utilities\") pod \"redhat-marketplace-j7pg7\" (UID: \"58792ed4-e360-4827-bf8c-508146205b78\") " pod="openshift-marketplace/redhat-marketplace-j7pg7" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.966244 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58792ed4-e360-4827-bf8c-508146205b78-catalog-content\") pod \"redhat-marketplace-j7pg7\" (UID: \"58792ed4-e360-4827-bf8c-508146205b78\") " pod="openshift-marketplace/redhat-marketplace-j7pg7" Dec 09 09:49:51 crc kubenswrapper[4824]: I1209 09:49:51.987081 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jz54p\" (UniqueName: \"kubernetes.io/projected/58792ed4-e360-4827-bf8c-508146205b78-kube-api-access-jz54p\") pod \"redhat-marketplace-j7pg7\" (UID: \"58792ed4-e360-4827-bf8c-508146205b78\") " pod="openshift-marketplace/redhat-marketplace-j7pg7" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.062473 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"7e74963449bbb47c83ec5a3f4d8b414f54b04d21cd3f11bd0aec59ac533be679"} Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.062855 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"76e3fe0dcc1522e50422a7b99213f8628c34794e39f950cabc892e5e4cf8d8e0"} Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.063696 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.066162 4824 generic.go:334] "Generic (PLEG): container finished" podID="ad53b3fe-0e9e-4d56-9ca0-212dde3ac669" containerID="6501aeec075b8a0c31f3c5dbc033f30ecdc3cba0674dde35b9bae12572492c80" exitCode=0 Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.066412 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xg5qv" event={"ID":"ad53b3fe-0e9e-4d56-9ca0-212dde3ac669","Type":"ContainerDied","Data":"6501aeec075b8a0c31f3c5dbc033f30ecdc3cba0674dde35b9bae12572492c80"} Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.067941 4824 generic.go:334] "Generic (PLEG): container finished" podID="e799b94c-33db-4de6-afd4-a8ffc4e9b069" containerID="f4043c535e6fcd55a4841f1ff3aa3e27e70824ff475563a214761e258a197dd6" exitCode=0 Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.067998 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zmxrv" event={"ID":"e799b94c-33db-4de6-afd4-a8ffc4e9b069","Type":"ContainerDied","Data":"f4043c535e6fcd55a4841f1ff3aa3e27e70824ff475563a214761e258a197dd6"} Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.069666 4824 generic.go:334] "Generic (PLEG): container finished" podID="741767ab-ed9f-4b68-83bb-41aa25497741" containerID="08e14507f9df56ef50a945ee72ab5c6c72f1ea4b9d3e82f51755ec5f98db9954" exitCode=0 Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.069769 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g9g6w" event={"ID":"741767ab-ed9f-4b68-83bb-41aa25497741","Type":"ContainerDied","Data":"08e14507f9df56ef50a945ee72ab5c6c72f1ea4b9d3e82f51755ec5f98db9954"} Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.071674 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"622dfa6d7e13379c434c2976b560e6300278c68f905ba6c136ceb5dfa533e294"} Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.071700 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"3230d69ff6f0b93927baf124b89a39b9f013e76ba9f10b8451852c81a15698a3"} Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.077268 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" event={"ID":"6d910d95-4961-4d28-9599-087153ce7d2f","Type":"ContainerStarted","Data":"d197cdd6496cac6eb5e8db0458314c964b45b0ebe8bf001ebc5b60eee8a75e47"} Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.077330 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" event={"ID":"6d910d95-4961-4d28-9599-087153ce7d2f","Type":"ContainerStarted","Data":"1b5175598a7d3f9dc435aec37623ff6fed9118de27a47b9b4509953362a3348a"} Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.078201 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.082640 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"ed6e450208dd84545603760b54984822ed3a0534cab9dc178bafeee9f943715e"} Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.082693 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"b73c6131f2328df15573a55c4aed52935b729e0d66a7351ae1fab6f69829eb90"} Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.094193 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.127943 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j7pg7" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.203720 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hrhmc"] Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.205174 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hrhmc" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.226714 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hrhmc"] Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.239152 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" podStartSLOduration=128.238971556 podStartE2EDuration="2m8.238971556s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:49:52.236211706 +0000 UTC m=+148.570716383" watchObservedRunningTime="2025-12-09 09:49:52.238971556 +0000 UTC m=+148.573476223" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.377210 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxsnt\" (UniqueName: \"kubernetes.io/projected/e43afd8e-2c79-454d-9c95-a0f65f13dc37-kube-api-access-bxsnt\") pod \"redhat-marketplace-hrhmc\" (UID: \"e43afd8e-2c79-454d-9c95-a0f65f13dc37\") " pod="openshift-marketplace/redhat-marketplace-hrhmc" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.377372 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e43afd8e-2c79-454d-9c95-a0f65f13dc37-utilities\") pod \"redhat-marketplace-hrhmc\" (UID: \"e43afd8e-2c79-454d-9c95-a0f65f13dc37\") " pod="openshift-marketplace/redhat-marketplace-hrhmc" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.377459 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e43afd8e-2c79-454d-9c95-a0f65f13dc37-catalog-content\") pod \"redhat-marketplace-hrhmc\" (UID: \"e43afd8e-2c79-454d-9c95-a0f65f13dc37\") " pod="openshift-marketplace/redhat-marketplace-hrhmc" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.479072 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e43afd8e-2c79-454d-9c95-a0f65f13dc37-catalog-content\") pod \"redhat-marketplace-hrhmc\" (UID: \"e43afd8e-2c79-454d-9c95-a0f65f13dc37\") " pod="openshift-marketplace/redhat-marketplace-hrhmc" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.479586 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxsnt\" (UniqueName: \"kubernetes.io/projected/e43afd8e-2c79-454d-9c95-a0f65f13dc37-kube-api-access-bxsnt\") pod \"redhat-marketplace-hrhmc\" (UID: \"e43afd8e-2c79-454d-9c95-a0f65f13dc37\") " pod="openshift-marketplace/redhat-marketplace-hrhmc" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.480040 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e43afd8e-2c79-454d-9c95-a0f65f13dc37-catalog-content\") pod \"redhat-marketplace-hrhmc\" (UID: \"e43afd8e-2c79-454d-9c95-a0f65f13dc37\") " pod="openshift-marketplace/redhat-marketplace-hrhmc" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.480276 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e43afd8e-2c79-454d-9c95-a0f65f13dc37-utilities\") pod \"redhat-marketplace-hrhmc\" (UID: \"e43afd8e-2c79-454d-9c95-a0f65f13dc37\") " pod="openshift-marketplace/redhat-marketplace-hrhmc" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.480859 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e43afd8e-2c79-454d-9c95-a0f65f13dc37-utilities\") pod \"redhat-marketplace-hrhmc\" (UID: \"e43afd8e-2c79-454d-9c95-a0f65f13dc37\") " pod="openshift-marketplace/redhat-marketplace-hrhmc" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.502005 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxsnt\" (UniqueName: \"kubernetes.io/projected/e43afd8e-2c79-454d-9c95-a0f65f13dc37-kube-api-access-bxsnt\") pod \"redhat-marketplace-hrhmc\" (UID: \"e43afd8e-2c79-454d-9c95-a0f65f13dc37\") " pod="openshift-marketplace/redhat-marketplace-hrhmc" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.571653 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hrhmc" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.666164 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j7pg7"] Dec 09 09:49:52 crc kubenswrapper[4824]: W1209 09:49:52.698897 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58792ed4_e360_4827_bf8c_508146205b78.slice/crio-a9e9ab1dd2ff7148955bea103166ef1a37504cb02d57cf81b810e0803afacd7e WatchSource:0}: Error finding container a9e9ab1dd2ff7148955bea103166ef1a37504cb02d57cf81b810e0803afacd7e: Status 404 returned error can't find the container with id a9e9ab1dd2ff7148955bea103166ef1a37504cb02d57cf81b810e0803afacd7e Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.794187 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qpbgg"] Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.795598 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qpbgg" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.795800 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qpbgg"] Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.798868 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.814158 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.814581 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.822328 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-b9bnh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.822365 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b9bnh" podUID="6efb8245-07e9-4c31-b8ab-53ab3685e593" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.822406 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-b9bnh container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.822523 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-b9bnh" podUID="6efb8245-07e9-4c31-b8ab-53ab3685e593" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.824443 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.836005 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.843332 4824 patch_prober.go:28] interesting pod/router-default-5444994796-mj7jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 09:49:52 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Dec 09 09:49:52 crc kubenswrapper[4824]: [+]process-running ok Dec 09 09:49:52 crc kubenswrapper[4824]: healthz check failed Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.843386 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.887428 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5qsr\" (UniqueName: \"kubernetes.io/projected/3715eab2-b516-4d39-a23f-664a70efe5b8-kube-api-access-b5qsr\") pod \"redhat-operators-qpbgg\" (UID: \"3715eab2-b516-4d39-a23f-664a70efe5b8\") " pod="openshift-marketplace/redhat-operators-qpbgg" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.887487 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3715eab2-b516-4d39-a23f-664a70efe5b8-utilities\") pod \"redhat-operators-qpbgg\" (UID: \"3715eab2-b516-4d39-a23f-664a70efe5b8\") " pod="openshift-marketplace/redhat-operators-qpbgg" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.887617 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3715eab2-b516-4d39-a23f-664a70efe5b8-catalog-content\") pod \"redhat-operators-qpbgg\" (UID: \"3715eab2-b516-4d39-a23f-664a70efe5b8\") " pod="openshift-marketplace/redhat-operators-qpbgg" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.989263 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3715eab2-b516-4d39-a23f-664a70efe5b8-catalog-content\") pod \"redhat-operators-qpbgg\" (UID: \"3715eab2-b516-4d39-a23f-664a70efe5b8\") " pod="openshift-marketplace/redhat-operators-qpbgg" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.989390 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5qsr\" (UniqueName: \"kubernetes.io/projected/3715eab2-b516-4d39-a23f-664a70efe5b8-kube-api-access-b5qsr\") pod \"redhat-operators-qpbgg\" (UID: \"3715eab2-b516-4d39-a23f-664a70efe5b8\") " pod="openshift-marketplace/redhat-operators-qpbgg" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.989425 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3715eab2-b516-4d39-a23f-664a70efe5b8-utilities\") pod \"redhat-operators-qpbgg\" (UID: \"3715eab2-b516-4d39-a23f-664a70efe5b8\") " pod="openshift-marketplace/redhat-operators-qpbgg" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.993387 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3715eab2-b516-4d39-a23f-664a70efe5b8-catalog-content\") pod \"redhat-operators-qpbgg\" (UID: \"3715eab2-b516-4d39-a23f-664a70efe5b8\") " pod="openshift-marketplace/redhat-operators-qpbgg" Dec 09 09:49:52 crc kubenswrapper[4824]: I1209 09:49:52.994657 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3715eab2-b516-4d39-a23f-664a70efe5b8-utilities\") pod \"redhat-operators-qpbgg\" (UID: \"3715eab2-b516-4d39-a23f-664a70efe5b8\") " pod="openshift-marketplace/redhat-operators-qpbgg" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.022924 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5qsr\" (UniqueName: \"kubernetes.io/projected/3715eab2-b516-4d39-a23f-664a70efe5b8-kube-api-access-b5qsr\") pod \"redhat-operators-qpbgg\" (UID: \"3715eab2-b516-4d39-a23f-664a70efe5b8\") " pod="openshift-marketplace/redhat-operators-qpbgg" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.084722 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hrhmc"] Dec 09 09:49:53 crc kubenswrapper[4824]: W1209 09:49:53.094439 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode43afd8e_2c79_454d_9c95_a0f65f13dc37.slice/crio-51928036c9ebedbb7d909a164d67cd07f5c9ec1bf4a78371bcf4ee1da4d87999 WatchSource:0}: Error finding container 51928036c9ebedbb7d909a164d67cd07f5c9ec1bf4a78371bcf4ee1da4d87999: Status 404 returned error can't find the container with id 51928036c9ebedbb7d909a164d67cd07f5c9ec1bf4a78371bcf4ee1da4d87999 Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.096095 4824 generic.go:334] "Generic (PLEG): container finished" podID="58792ed4-e360-4827-bf8c-508146205b78" containerID="df6bde8f2b4593a55bca46d073d569cdf13cae123f4bc5d63c2e2932a77affe5" exitCode=0 Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.096174 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7pg7" event={"ID":"58792ed4-e360-4827-bf8c-508146205b78","Type":"ContainerDied","Data":"df6bde8f2b4593a55bca46d073d569cdf13cae123f4bc5d63c2e2932a77affe5"} Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.096210 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7pg7" event={"ID":"58792ed4-e360-4827-bf8c-508146205b78","Type":"ContainerStarted","Data":"a9e9ab1dd2ff7148955bea103166ef1a37504cb02d57cf81b810e0803afacd7e"} Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.103206 4824 generic.go:334] "Generic (PLEG): container finished" podID="055ead87-3e57-4053-a026-24566879437c" containerID="91eeeddf7380a1f76b21a29c04f33b21a23c21087506d6559bdb21b8ad896cac" exitCode=0 Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.103847 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421225-96fgb" event={"ID":"055ead87-3e57-4053-a026-24566879437c","Type":"ContainerDied","Data":"91eeeddf7380a1f76b21a29c04f33b21a23c21087506d6559bdb21b8ad896cac"} Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.113076 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-pd82d" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.136126 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qpbgg" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.211710 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4swl5"] Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.221345 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4swl5" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.228090 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4swl5"] Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.372732 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.373922 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.377316 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.377313 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.379310 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.399587 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj9rd\" (UniqueName: \"kubernetes.io/projected/7a6c3ca8-ba36-4cbc-9143-4f83a611144a-kube-api-access-nj9rd\") pod \"redhat-operators-4swl5\" (UID: \"7a6c3ca8-ba36-4cbc-9143-4f83a611144a\") " pod="openshift-marketplace/redhat-operators-4swl5" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.399668 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6c3ca8-ba36-4cbc-9143-4f83a611144a-utilities\") pod \"redhat-operators-4swl5\" (UID: \"7a6c3ca8-ba36-4cbc-9143-4f83a611144a\") " pod="openshift-marketplace/redhat-operators-4swl5" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.399703 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6c3ca8-ba36-4cbc-9143-4f83a611144a-catalog-content\") pod \"redhat-operators-4swl5\" (UID: \"7a6c3ca8-ba36-4cbc-9143-4f83a611144a\") " pod="openshift-marketplace/redhat-operators-4swl5" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.510493 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3fbd2041-d53d-4ccd-8758-cb177f3d7eea-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3fbd2041-d53d-4ccd-8758-cb177f3d7eea\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.510571 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3fbd2041-d53d-4ccd-8758-cb177f3d7eea-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3fbd2041-d53d-4ccd-8758-cb177f3d7eea\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.510622 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj9rd\" (UniqueName: \"kubernetes.io/projected/7a6c3ca8-ba36-4cbc-9143-4f83a611144a-kube-api-access-nj9rd\") pod \"redhat-operators-4swl5\" (UID: \"7a6c3ca8-ba36-4cbc-9143-4f83a611144a\") " pod="openshift-marketplace/redhat-operators-4swl5" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.510662 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6c3ca8-ba36-4cbc-9143-4f83a611144a-utilities\") pod \"redhat-operators-4swl5\" (UID: \"7a6c3ca8-ba36-4cbc-9143-4f83a611144a\") " pod="openshift-marketplace/redhat-operators-4swl5" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.510694 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6c3ca8-ba36-4cbc-9143-4f83a611144a-catalog-content\") pod \"redhat-operators-4swl5\" (UID: \"7a6c3ca8-ba36-4cbc-9143-4f83a611144a\") " pod="openshift-marketplace/redhat-operators-4swl5" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.511235 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6c3ca8-ba36-4cbc-9143-4f83a611144a-catalog-content\") pod \"redhat-operators-4swl5\" (UID: \"7a6c3ca8-ba36-4cbc-9143-4f83a611144a\") " pod="openshift-marketplace/redhat-operators-4swl5" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.511827 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6c3ca8-ba36-4cbc-9143-4f83a611144a-utilities\") pod \"redhat-operators-4swl5\" (UID: \"7a6c3ca8-ba36-4cbc-9143-4f83a611144a\") " pod="openshift-marketplace/redhat-operators-4swl5" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.539466 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj9rd\" (UniqueName: \"kubernetes.io/projected/7a6c3ca8-ba36-4cbc-9143-4f83a611144a-kube-api-access-nj9rd\") pod \"redhat-operators-4swl5\" (UID: \"7a6c3ca8-ba36-4cbc-9143-4f83a611144a\") " pod="openshift-marketplace/redhat-operators-4swl5" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.539891 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4swl5" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.558080 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qpbgg"] Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.612084 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3fbd2041-d53d-4ccd-8758-cb177f3d7eea-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3fbd2041-d53d-4ccd-8758-cb177f3d7eea\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.612509 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3fbd2041-d53d-4ccd-8758-cb177f3d7eea-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3fbd2041-d53d-4ccd-8758-cb177f3d7eea\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.612910 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3fbd2041-d53d-4ccd-8758-cb177f3d7eea-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3fbd2041-d53d-4ccd-8758-cb177f3d7eea\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.688308 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3fbd2041-d53d-4ccd-8758-cb177f3d7eea-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3fbd2041-d53d-4ccd-8758-cb177f3d7eea\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.689818 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.691013 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.692824 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.696727 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.701065 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.702059 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.815837 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8793f797-234b-463b-a859-0e5f55d5f7a3-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"8793f797-234b-463b-a859-0e5f55d5f7a3\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.816071 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8793f797-234b-463b-a859-0e5f55d5f7a3-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"8793f797-234b-463b-a859-0e5f55d5f7a3\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.842804 4824 patch_prober.go:28] interesting pod/router-default-5444994796-mj7jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 09:49:53 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Dec 09 09:49:53 crc kubenswrapper[4824]: [+]process-running ok Dec 09 09:49:53 crc kubenswrapper[4824]: healthz check failed Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.842880 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.919140 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8793f797-234b-463b-a859-0e5f55d5f7a3-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"8793f797-234b-463b-a859-0e5f55d5f7a3\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.919624 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8793f797-234b-463b-a859-0e5f55d5f7a3-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"8793f797-234b-463b-a859-0e5f55d5f7a3\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.919819 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8793f797-234b-463b-a859-0e5f55d5f7a3-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"8793f797-234b-463b-a859-0e5f55d5f7a3\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 09:49:53 crc kubenswrapper[4824]: I1209 09:49:53.980623 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8793f797-234b-463b-a859-0e5f55d5f7a3-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"8793f797-234b-463b-a859-0e5f55d5f7a3\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 09:49:54 crc kubenswrapper[4824]: I1209 09:49:54.019666 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 09:49:54 crc kubenswrapper[4824]: I1209 09:49:54.173360 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 09 09:49:54 crc kubenswrapper[4824]: I1209 09:49:54.189911 4824 generic.go:334] "Generic (PLEG): container finished" podID="3715eab2-b516-4d39-a23f-664a70efe5b8" containerID="a938c67634e4808db60c1728f3a58403f72fdb7f5f003adf3973cb3ed2dc52c2" exitCode=0 Dec 09 09:49:54 crc kubenswrapper[4824]: I1209 09:49:54.190437 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qpbgg" event={"ID":"3715eab2-b516-4d39-a23f-664a70efe5b8","Type":"ContainerDied","Data":"a938c67634e4808db60c1728f3a58403f72fdb7f5f003adf3973cb3ed2dc52c2"} Dec 09 09:49:54 crc kubenswrapper[4824]: I1209 09:49:54.192276 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qpbgg" event={"ID":"3715eab2-b516-4d39-a23f-664a70efe5b8","Type":"ContainerStarted","Data":"96955b17111cd0bbe43ca287387c357b0877f0951747ea335b834b8749f93d43"} Dec 09 09:49:54 crc kubenswrapper[4824]: I1209 09:49:54.228299 4824 generic.go:334] "Generic (PLEG): container finished" podID="e43afd8e-2c79-454d-9c95-a0f65f13dc37" containerID="352c1aeab5048b8931929b28166d07fabb6fa498f223e192d9b3ffcbeb23603a" exitCode=0 Dec 09 09:49:54 crc kubenswrapper[4824]: I1209 09:49:54.229505 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hrhmc" event={"ID":"e43afd8e-2c79-454d-9c95-a0f65f13dc37","Type":"ContainerDied","Data":"352c1aeab5048b8931929b28166d07fabb6fa498f223e192d9b3ffcbeb23603a"} Dec 09 09:49:54 crc kubenswrapper[4824]: I1209 09:49:54.229531 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hrhmc" event={"ID":"e43afd8e-2c79-454d-9c95-a0f65f13dc37","Type":"ContainerStarted","Data":"51928036c9ebedbb7d909a164d67cd07f5c9ec1bf4a78371bcf4ee1da4d87999"} Dec 09 09:49:54 crc kubenswrapper[4824]: I1209 09:49:54.306277 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4swl5"] Dec 09 09:49:54 crc kubenswrapper[4824]: W1209 09:49:54.389542 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a6c3ca8_ba36_4cbc_9143_4f83a611144a.slice/crio-b247c8841ba6392effdc3604929acb2447df14980fb04bc3c8e77f03a3062f2f WatchSource:0}: Error finding container b247c8841ba6392effdc3604929acb2447df14980fb04bc3c8e77f03a3062f2f: Status 404 returned error can't find the container with id b247c8841ba6392effdc3604929acb2447df14980fb04bc3c8e77f03a3062f2f Dec 09 09:49:54 crc kubenswrapper[4824]: I1209 09:49:54.450580 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 09 09:49:54 crc kubenswrapper[4824]: W1209 09:49:54.489878 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod8793f797_234b_463b_a859_0e5f55d5f7a3.slice/crio-38943d62af870bc6685cfdb1fa28c0d50cc50759d686af9c6a256a577bd05ca9 WatchSource:0}: Error finding container 38943d62af870bc6685cfdb1fa28c0d50cc50759d686af9c6a256a577bd05ca9: Status 404 returned error can't find the container with id 38943d62af870bc6685cfdb1fa28c0d50cc50759d686af9c6a256a577bd05ca9 Dec 09 09:49:54 crc kubenswrapper[4824]: I1209 09:49:54.641479 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:49:54 crc kubenswrapper[4824]: I1209 09:49:54.763447 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421225-96fgb" Dec 09 09:49:54 crc kubenswrapper[4824]: I1209 09:49:54.840632 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qwml\" (UniqueName: \"kubernetes.io/projected/055ead87-3e57-4053-a026-24566879437c-kube-api-access-2qwml\") pod \"055ead87-3e57-4053-a026-24566879437c\" (UID: \"055ead87-3e57-4053-a026-24566879437c\") " Dec 09 09:49:54 crc kubenswrapper[4824]: I1209 09:49:54.840710 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/055ead87-3e57-4053-a026-24566879437c-config-volume\") pod \"055ead87-3e57-4053-a026-24566879437c\" (UID: \"055ead87-3e57-4053-a026-24566879437c\") " Dec 09 09:49:54 crc kubenswrapper[4824]: I1209 09:49:54.840800 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/055ead87-3e57-4053-a026-24566879437c-secret-volume\") pod \"055ead87-3e57-4053-a026-24566879437c\" (UID: \"055ead87-3e57-4053-a026-24566879437c\") " Dec 09 09:49:54 crc kubenswrapper[4824]: I1209 09:49:54.842217 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/055ead87-3e57-4053-a026-24566879437c-config-volume" (OuterVolumeSpecName: "config-volume") pod "055ead87-3e57-4053-a026-24566879437c" (UID: "055ead87-3e57-4053-a026-24566879437c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:49:54 crc kubenswrapper[4824]: I1209 09:49:54.846749 4824 patch_prober.go:28] interesting pod/router-default-5444994796-mj7jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 09:49:54 crc kubenswrapper[4824]: [+]has-synced ok Dec 09 09:49:54 crc kubenswrapper[4824]: [+]process-running ok Dec 09 09:49:54 crc kubenswrapper[4824]: healthz check failed Dec 09 09:49:54 crc kubenswrapper[4824]: I1209 09:49:54.846825 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 09:49:54 crc kubenswrapper[4824]: I1209 09:49:54.855977 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/055ead87-3e57-4053-a026-24566879437c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "055ead87-3e57-4053-a026-24566879437c" (UID: "055ead87-3e57-4053-a026-24566879437c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:49:54 crc kubenswrapper[4824]: I1209 09:49:54.877752 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/055ead87-3e57-4053-a026-24566879437c-kube-api-access-2qwml" (OuterVolumeSpecName: "kube-api-access-2qwml") pod "055ead87-3e57-4053-a026-24566879437c" (UID: "055ead87-3e57-4053-a026-24566879437c"). InnerVolumeSpecName "kube-api-access-2qwml". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:49:54 crc kubenswrapper[4824]: I1209 09:49:54.941889 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qwml\" (UniqueName: \"kubernetes.io/projected/055ead87-3e57-4053-a026-24566879437c-kube-api-access-2qwml\") on node \"crc\" DevicePath \"\"" Dec 09 09:49:54 crc kubenswrapper[4824]: I1209 09:49:54.941923 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/055ead87-3e57-4053-a026-24566879437c-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 09:49:54 crc kubenswrapper[4824]: I1209 09:49:54.941935 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/055ead87-3e57-4053-a026-24566879437c-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 09:49:55 crc kubenswrapper[4824]: I1209 09:49:55.290551 4824 generic.go:334] "Generic (PLEG): container finished" podID="7a6c3ca8-ba36-4cbc-9143-4f83a611144a" containerID="fb95a63af9504846fdcf803d4dd2675bdbb03544b6030a50c4a146cf65b5448a" exitCode=0 Dec 09 09:49:55 crc kubenswrapper[4824]: I1209 09:49:55.290911 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4swl5" event={"ID":"7a6c3ca8-ba36-4cbc-9143-4f83a611144a","Type":"ContainerDied","Data":"fb95a63af9504846fdcf803d4dd2675bdbb03544b6030a50c4a146cf65b5448a"} Dec 09 09:49:55 crc kubenswrapper[4824]: I1209 09:49:55.291459 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4swl5" event={"ID":"7a6c3ca8-ba36-4cbc-9143-4f83a611144a","Type":"ContainerStarted","Data":"b247c8841ba6392effdc3604929acb2447df14980fb04bc3c8e77f03a3062f2f"} Dec 09 09:49:55 crc kubenswrapper[4824]: I1209 09:49:55.342976 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421225-96fgb" event={"ID":"055ead87-3e57-4053-a026-24566879437c","Type":"ContainerDied","Data":"636e96bbd2a46a602b57695d044dc991d80616bee0bacea2972ec7bba0e16bff"} Dec 09 09:49:55 crc kubenswrapper[4824]: I1209 09:49:55.343001 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421225-96fgb" Dec 09 09:49:55 crc kubenswrapper[4824]: I1209 09:49:55.343015 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="636e96bbd2a46a602b57695d044dc991d80616bee0bacea2972ec7bba0e16bff" Dec 09 09:49:55 crc kubenswrapper[4824]: I1209 09:49:55.349969 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"8793f797-234b-463b-a859-0e5f55d5f7a3","Type":"ContainerStarted","Data":"38943d62af870bc6685cfdb1fa28c0d50cc50759d686af9c6a256a577bd05ca9"} Dec 09 09:49:55 crc kubenswrapper[4824]: I1209 09:49:55.376316 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3fbd2041-d53d-4ccd-8758-cb177f3d7eea","Type":"ContainerStarted","Data":"9196569bd963c305dfcdb3b73c50c579379969814101c411fbf4d22a85c65ecb"} Dec 09 09:49:55 crc kubenswrapper[4824]: I1209 09:49:55.845657 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 09:49:55 crc kubenswrapper[4824]: I1209 09:49:55.853597 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 09:49:56 crc kubenswrapper[4824]: I1209 09:49:56.403122 4824 generic.go:334] "Generic (PLEG): container finished" podID="3fbd2041-d53d-4ccd-8758-cb177f3d7eea" containerID="245f300472c64bab25fb4302c70dc5d9990d0f228f4e590a1ae58331651d9b18" exitCode=0 Dec 09 09:49:56 crc kubenswrapper[4824]: I1209 09:49:56.403357 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3fbd2041-d53d-4ccd-8758-cb177f3d7eea","Type":"ContainerDied","Data":"245f300472c64bab25fb4302c70dc5d9990d0f228f4e590a1ae58331651d9b18"} Dec 09 09:49:56 crc kubenswrapper[4824]: I1209 09:49:56.408093 4824 generic.go:334] "Generic (PLEG): container finished" podID="8793f797-234b-463b-a859-0e5f55d5f7a3" containerID="4dad7f6e5f3dba5e5af851ada661274e340ca5e49a09b36d59b4c93104ab10f2" exitCode=0 Dec 09 09:49:56 crc kubenswrapper[4824]: I1209 09:49:56.408971 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"8793f797-234b-463b-a859-0e5f55d5f7a3","Type":"ContainerDied","Data":"4dad7f6e5f3dba5e5af851ada661274e340ca5e49a09b36d59b4c93104ab10f2"} Dec 09 09:49:57 crc kubenswrapper[4824]: I1209 09:49:57.836060 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 09:49:57 crc kubenswrapper[4824]: I1209 09:49:57.956364 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 09:49:58 crc kubenswrapper[4824]: I1209 09:49:58.013755 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-zjjp6" Dec 09 09:49:58 crc kubenswrapper[4824]: I1209 09:49:58.026864 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3fbd2041-d53d-4ccd-8758-cb177f3d7eea-kube-api-access\") pod \"3fbd2041-d53d-4ccd-8758-cb177f3d7eea\" (UID: \"3fbd2041-d53d-4ccd-8758-cb177f3d7eea\") " Dec 09 09:49:58 crc kubenswrapper[4824]: I1209 09:49:58.026926 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3fbd2041-d53d-4ccd-8758-cb177f3d7eea-kubelet-dir\") pod \"3fbd2041-d53d-4ccd-8758-cb177f3d7eea\" (UID: \"3fbd2041-d53d-4ccd-8758-cb177f3d7eea\") " Dec 09 09:49:58 crc kubenswrapper[4824]: I1209 09:49:58.027282 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3fbd2041-d53d-4ccd-8758-cb177f3d7eea-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3fbd2041-d53d-4ccd-8758-cb177f3d7eea" (UID: "3fbd2041-d53d-4ccd-8758-cb177f3d7eea"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:49:58 crc kubenswrapper[4824]: I1209 09:49:58.048268 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fbd2041-d53d-4ccd-8758-cb177f3d7eea-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3fbd2041-d53d-4ccd-8758-cb177f3d7eea" (UID: "3fbd2041-d53d-4ccd-8758-cb177f3d7eea"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:49:58 crc kubenswrapper[4824]: I1209 09:49:58.129314 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8793f797-234b-463b-a859-0e5f55d5f7a3-kube-api-access\") pod \"8793f797-234b-463b-a859-0e5f55d5f7a3\" (UID: \"8793f797-234b-463b-a859-0e5f55d5f7a3\") " Dec 09 09:49:58 crc kubenswrapper[4824]: I1209 09:49:58.129473 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8793f797-234b-463b-a859-0e5f55d5f7a3-kubelet-dir\") pod \"8793f797-234b-463b-a859-0e5f55d5f7a3\" (UID: \"8793f797-234b-463b-a859-0e5f55d5f7a3\") " Dec 09 09:49:58 crc kubenswrapper[4824]: I1209 09:49:58.129622 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8793f797-234b-463b-a859-0e5f55d5f7a3-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "8793f797-234b-463b-a859-0e5f55d5f7a3" (UID: "8793f797-234b-463b-a859-0e5f55d5f7a3"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:49:58 crc kubenswrapper[4824]: I1209 09:49:58.130040 4824 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8793f797-234b-463b-a859-0e5f55d5f7a3-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 09 09:49:58 crc kubenswrapper[4824]: I1209 09:49:58.130066 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3fbd2041-d53d-4ccd-8758-cb177f3d7eea-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 09:49:58 crc kubenswrapper[4824]: I1209 09:49:58.130085 4824 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3fbd2041-d53d-4ccd-8758-cb177f3d7eea-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 09 09:49:58 crc kubenswrapper[4824]: I1209 09:49:58.135293 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8793f797-234b-463b-a859-0e5f55d5f7a3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "8793f797-234b-463b-a859-0e5f55d5f7a3" (UID: "8793f797-234b-463b-a859-0e5f55d5f7a3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:49:58 crc kubenswrapper[4824]: I1209 09:49:58.231414 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8793f797-234b-463b-a859-0e5f55d5f7a3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 09:49:58 crc kubenswrapper[4824]: I1209 09:49:58.442046 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 09:49:58 crc kubenswrapper[4824]: I1209 09:49:58.442052 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3fbd2041-d53d-4ccd-8758-cb177f3d7eea","Type":"ContainerDied","Data":"9196569bd963c305dfcdb3b73c50c579379969814101c411fbf4d22a85c65ecb"} Dec 09 09:49:58 crc kubenswrapper[4824]: I1209 09:49:58.442090 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9196569bd963c305dfcdb3b73c50c579379969814101c411fbf4d22a85c65ecb" Dec 09 09:49:58 crc kubenswrapper[4824]: I1209 09:49:58.454231 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"8793f797-234b-463b-a859-0e5f55d5f7a3","Type":"ContainerDied","Data":"38943d62af870bc6685cfdb1fa28c0d50cc50759d686af9c6a256a577bd05ca9"} Dec 09 09:49:58 crc kubenswrapper[4824]: I1209 09:49:58.454306 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38943d62af870bc6685cfdb1fa28c0d50cc50759d686af9c6a256a577bd05ca9" Dec 09 09:49:58 crc kubenswrapper[4824]: I1209 09:49:58.454367 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 09:50:01 crc kubenswrapper[4824]: I1209 09:50:01.924201 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:50:01 crc kubenswrapper[4824]: I1209 09:50:01.928709 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:50:02 crc kubenswrapper[4824]: I1209 09:50:02.828435 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-b9bnh" Dec 09 09:50:02 crc kubenswrapper[4824]: I1209 09:50:02.860923 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 09:50:02 crc kubenswrapper[4824]: I1209 09:50:02.861023 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 09:50:06 crc kubenswrapper[4824]: I1209 09:50:06.536641 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs\") pod \"network-metrics-daemon-5tn52\" (UID: \"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\") " pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:50:06 crc kubenswrapper[4824]: I1209 09:50:06.560351 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcbce9ea-3d55-4e09-b5bc-d5bb8c472219-metrics-certs\") pod \"network-metrics-daemon-5tn52\" (UID: \"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219\") " pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:50:06 crc kubenswrapper[4824]: I1209 09:50:06.747694 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5tn52" Dec 09 09:50:11 crc kubenswrapper[4824]: I1209 09:50:11.484160 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:50:21 crc kubenswrapper[4824]: I1209 09:50:21.198729 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 09:50:22 crc kubenswrapper[4824]: I1209 09:50:22.979322 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" Dec 09 09:50:27 crc kubenswrapper[4824]: I1209 09:50:27.821679 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 09 09:50:27 crc kubenswrapper[4824]: E1209 09:50:27.823916 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8793f797-234b-463b-a859-0e5f55d5f7a3" containerName="pruner" Dec 09 09:50:27 crc kubenswrapper[4824]: I1209 09:50:27.824042 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8793f797-234b-463b-a859-0e5f55d5f7a3" containerName="pruner" Dec 09 09:50:27 crc kubenswrapper[4824]: E1209 09:50:27.824173 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="055ead87-3e57-4053-a026-24566879437c" containerName="collect-profiles" Dec 09 09:50:27 crc kubenswrapper[4824]: I1209 09:50:27.824277 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="055ead87-3e57-4053-a026-24566879437c" containerName="collect-profiles" Dec 09 09:50:27 crc kubenswrapper[4824]: E1209 09:50:27.824370 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fbd2041-d53d-4ccd-8758-cb177f3d7eea" containerName="pruner" Dec 09 09:50:27 crc kubenswrapper[4824]: I1209 09:50:27.824464 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fbd2041-d53d-4ccd-8758-cb177f3d7eea" containerName="pruner" Dec 09 09:50:27 crc kubenswrapper[4824]: I1209 09:50:27.824674 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fbd2041-d53d-4ccd-8758-cb177f3d7eea" containerName="pruner" Dec 09 09:50:27 crc kubenswrapper[4824]: I1209 09:50:27.824776 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8793f797-234b-463b-a859-0e5f55d5f7a3" containerName="pruner" Dec 09 09:50:27 crc kubenswrapper[4824]: I1209 09:50:27.825153 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="055ead87-3e57-4053-a026-24566879437c" containerName="collect-profiles" Dec 09 09:50:27 crc kubenswrapper[4824]: I1209 09:50:27.825753 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 09:50:27 crc kubenswrapper[4824]: I1209 09:50:27.829227 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 09 09:50:27 crc kubenswrapper[4824]: I1209 09:50:27.829238 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 09 09:50:27 crc kubenswrapper[4824]: I1209 09:50:27.829998 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 09 09:50:28 crc kubenswrapper[4824]: I1209 09:50:28.165402 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0d7422fc-b0b4-4118-a0de-980eaf4f4fad-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0d7422fc-b0b4-4118-a0de-980eaf4f4fad\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 09:50:28 crc kubenswrapper[4824]: I1209 09:50:28.165758 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0d7422fc-b0b4-4118-a0de-980eaf4f4fad-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0d7422fc-b0b4-4118-a0de-980eaf4f4fad\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 09:50:28 crc kubenswrapper[4824]: I1209 09:50:28.267696 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0d7422fc-b0b4-4118-a0de-980eaf4f4fad-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0d7422fc-b0b4-4118-a0de-980eaf4f4fad\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 09:50:28 crc kubenswrapper[4824]: I1209 09:50:28.267816 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0d7422fc-b0b4-4118-a0de-980eaf4f4fad-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0d7422fc-b0b4-4118-a0de-980eaf4f4fad\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 09:50:28 crc kubenswrapper[4824]: I1209 09:50:28.267913 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0d7422fc-b0b4-4118-a0de-980eaf4f4fad-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0d7422fc-b0b4-4118-a0de-980eaf4f4fad\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 09:50:28 crc kubenswrapper[4824]: I1209 09:50:28.290976 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0d7422fc-b0b4-4118-a0de-980eaf4f4fad-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0d7422fc-b0b4-4118-a0de-980eaf4f4fad\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 09:50:28 crc kubenswrapper[4824]: I1209 09:50:28.458859 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 09:50:31 crc kubenswrapper[4824]: E1209 09:50:31.645151 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 09 09:50:31 crc kubenswrapper[4824]: E1209 09:50:31.646317 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qx6hv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-xg5qv_openshift-marketplace(ad53b3fe-0e9e-4d56-9ca0-212dde3ac669): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 09:50:31 crc kubenswrapper[4824]: E1209 09:50:31.647749 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-xg5qv" podUID="ad53b3fe-0e9e-4d56-9ca0-212dde3ac669" Dec 09 09:50:31 crc kubenswrapper[4824]: E1209 09:50:31.665761 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 09 09:50:31 crc kubenswrapper[4824]: E1209 09:50:31.666055 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-28b6h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-zmxrv_openshift-marketplace(e799b94c-33db-4de6-afd4-a8ffc4e9b069): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 09:50:31 crc kubenswrapper[4824]: E1209 09:50:31.667315 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-zmxrv" podUID="e799b94c-33db-4de6-afd4-a8ffc4e9b069" Dec 09 09:50:32 crc kubenswrapper[4824]: I1209 09:50:32.861617 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 09:50:32 crc kubenswrapper[4824]: I1209 09:50:32.862113 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 09:50:33 crc kubenswrapper[4824]: I1209 09:50:33.009991 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 09 09:50:33 crc kubenswrapper[4824]: I1209 09:50:33.014273 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 09 09:50:33 crc kubenswrapper[4824]: I1209 09:50:33.022175 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 09 09:50:33 crc kubenswrapper[4824]: I1209 09:50:33.184068 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/332c50f5-f531-4dc6-9f0f-9d903013feff-var-lock\") pod \"installer-9-crc\" (UID: \"332c50f5-f531-4dc6-9f0f-9d903013feff\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 09:50:33 crc kubenswrapper[4824]: I1209 09:50:33.185509 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/332c50f5-f531-4dc6-9f0f-9d903013feff-kubelet-dir\") pod \"installer-9-crc\" (UID: \"332c50f5-f531-4dc6-9f0f-9d903013feff\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 09:50:33 crc kubenswrapper[4824]: I1209 09:50:33.185641 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/332c50f5-f531-4dc6-9f0f-9d903013feff-kube-api-access\") pod \"installer-9-crc\" (UID: \"332c50f5-f531-4dc6-9f0f-9d903013feff\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 09:50:33 crc kubenswrapper[4824]: I1209 09:50:33.286642 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/332c50f5-f531-4dc6-9f0f-9d903013feff-kube-api-access\") pod \"installer-9-crc\" (UID: \"332c50f5-f531-4dc6-9f0f-9d903013feff\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 09:50:33 crc kubenswrapper[4824]: I1209 09:50:33.286723 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/332c50f5-f531-4dc6-9f0f-9d903013feff-var-lock\") pod \"installer-9-crc\" (UID: \"332c50f5-f531-4dc6-9f0f-9d903013feff\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 09:50:33 crc kubenswrapper[4824]: I1209 09:50:33.286747 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/332c50f5-f531-4dc6-9f0f-9d903013feff-kubelet-dir\") pod \"installer-9-crc\" (UID: \"332c50f5-f531-4dc6-9f0f-9d903013feff\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 09:50:33 crc kubenswrapper[4824]: I1209 09:50:33.286835 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/332c50f5-f531-4dc6-9f0f-9d903013feff-kubelet-dir\") pod \"installer-9-crc\" (UID: \"332c50f5-f531-4dc6-9f0f-9d903013feff\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 09:50:33 crc kubenswrapper[4824]: I1209 09:50:33.286920 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/332c50f5-f531-4dc6-9f0f-9d903013feff-var-lock\") pod \"installer-9-crc\" (UID: \"332c50f5-f531-4dc6-9f0f-9d903013feff\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 09:50:33 crc kubenswrapper[4824]: I1209 09:50:33.321321 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/332c50f5-f531-4dc6-9f0f-9d903013feff-kube-api-access\") pod \"installer-9-crc\" (UID: \"332c50f5-f531-4dc6-9f0f-9d903013feff\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 09:50:33 crc kubenswrapper[4824]: I1209 09:50:33.353295 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 09 09:50:35 crc kubenswrapper[4824]: E1209 09:50:35.631088 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xg5qv" podUID="ad53b3fe-0e9e-4d56-9ca0-212dde3ac669" Dec 09 09:50:35 crc kubenswrapper[4824]: E1209 09:50:35.631105 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-zmxrv" podUID="e799b94c-33db-4de6-afd4-a8ffc4e9b069" Dec 09 09:50:35 crc kubenswrapper[4824]: E1209 09:50:35.746112 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 09 09:50:35 crc kubenswrapper[4824]: E1209 09:50:35.746576 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nj9rd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-4swl5_openshift-marketplace(7a6c3ca8-ba36-4cbc-9143-4f83a611144a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 09:50:35 crc kubenswrapper[4824]: E1209 09:50:35.748287 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-4swl5" podUID="7a6c3ca8-ba36-4cbc-9143-4f83a611144a" Dec 09 09:50:37 crc kubenswrapper[4824]: E1209 09:50:37.319318 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-4swl5" podUID="7a6c3ca8-ba36-4cbc-9143-4f83a611144a" Dec 09 09:50:37 crc kubenswrapper[4824]: E1209 09:50:37.392091 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 09 09:50:37 crc kubenswrapper[4824]: E1209 09:50:37.392327 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vbkw6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-9pdnx_openshift-marketplace(db041858-ed93-4df9-a5d2-f79b7f9c3275): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 09:50:37 crc kubenswrapper[4824]: E1209 09:50:37.393917 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-9pdnx" podUID="db041858-ed93-4df9-a5d2-f79b7f9c3275" Dec 09 09:50:39 crc kubenswrapper[4824]: E1209 09:50:39.772840 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-9pdnx" podUID="db041858-ed93-4df9-a5d2-f79b7f9c3275" Dec 09 09:50:39 crc kubenswrapper[4824]: E1209 09:50:39.863668 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 09 09:50:39 crc kubenswrapper[4824]: E1209 09:50:39.863956 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bxsnt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-hrhmc_openshift-marketplace(e43afd8e-2c79-454d-9c95-a0f65f13dc37): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 09:50:39 crc kubenswrapper[4824]: E1209 09:50:39.865401 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-hrhmc" podUID="e43afd8e-2c79-454d-9c95-a0f65f13dc37" Dec 09 09:50:39 crc kubenswrapper[4824]: E1209 09:50:39.883215 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 09 09:50:39 crc kubenswrapper[4824]: E1209 09:50:39.883503 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jz54p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-j7pg7_openshift-marketplace(58792ed4-e360-4827-bf8c-508146205b78): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 09:50:39 crc kubenswrapper[4824]: E1209 09:50:39.884762 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-j7pg7" podUID="58792ed4-e360-4827-bf8c-508146205b78" Dec 09 09:50:39 crc kubenswrapper[4824]: E1209 09:50:39.998852 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 09 09:50:39 crc kubenswrapper[4824]: E1209 09:50:39.999614 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wswd5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-g9g6w_openshift-marketplace(741767ab-ed9f-4b68-83bb-41aa25497741): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 09:50:40 crc kubenswrapper[4824]: E1209 09:50:40.000881 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-g9g6w" podUID="741767ab-ed9f-4b68-83bb-41aa25497741" Dec 09 09:50:40 crc kubenswrapper[4824]: E1209 09:50:40.008620 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 09 09:50:40 crc kubenswrapper[4824]: E1209 09:50:40.009236 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b5qsr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-qpbgg_openshift-marketplace(3715eab2-b516-4d39-a23f-664a70efe5b8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 09:50:40 crc kubenswrapper[4824]: E1209 09:50:40.010373 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-qpbgg" podUID="3715eab2-b516-4d39-a23f-664a70efe5b8" Dec 09 09:50:40 crc kubenswrapper[4824]: I1209 09:50:40.254766 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-5tn52"] Dec 09 09:50:40 crc kubenswrapper[4824]: W1209 09:50:40.262809 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfcbce9ea_3d55_4e09_b5bc_d5bb8c472219.slice/crio-1b894aefaaebc57e7d0b8f268113fd7f180d94ccee69c8e6ba8e138edf1e899a WatchSource:0}: Error finding container 1b894aefaaebc57e7d0b8f268113fd7f180d94ccee69c8e6ba8e138edf1e899a: Status 404 returned error can't find the container with id 1b894aefaaebc57e7d0b8f268113fd7f180d94ccee69c8e6ba8e138edf1e899a Dec 09 09:50:40 crc kubenswrapper[4824]: E1209 09:50:40.269680 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-qpbgg" podUID="3715eab2-b516-4d39-a23f-664a70efe5b8" Dec 09 09:50:40 crc kubenswrapper[4824]: E1209 09:50:40.269854 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-g9g6w" podUID="741767ab-ed9f-4b68-83bb-41aa25497741" Dec 09 09:50:40 crc kubenswrapper[4824]: E1209 09:50:40.270414 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-hrhmc" podUID="e43afd8e-2c79-454d-9c95-a0f65f13dc37" Dec 09 09:50:40 crc kubenswrapper[4824]: E1209 09:50:40.270629 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-j7pg7" podUID="58792ed4-e360-4827-bf8c-508146205b78" Dec 09 09:50:40 crc kubenswrapper[4824]: I1209 09:50:40.335616 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 09 09:50:40 crc kubenswrapper[4824]: I1209 09:50:40.365355 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 09 09:50:41 crc kubenswrapper[4824]: I1209 09:50:41.272721 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"332c50f5-f531-4dc6-9f0f-9d903013feff","Type":"ContainerStarted","Data":"956fcedd182c664780197779f6d8b8c482bd461e33d1c2f0145e5b98a8b3e053"} Dec 09 09:50:41 crc kubenswrapper[4824]: I1209 09:50:41.273233 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"332c50f5-f531-4dc6-9f0f-9d903013feff","Type":"ContainerStarted","Data":"d2de8967b2631ed5500dcb843ff5db9cf56e3b0c7c4d4e4419df468e93096ef1"} Dec 09 09:50:41 crc kubenswrapper[4824]: I1209 09:50:41.275629 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"0d7422fc-b0b4-4118-a0de-980eaf4f4fad","Type":"ContainerStarted","Data":"70c9ba842d3301c8024d030476c0b5efc53c125d19cebc3cb2dc5716d21f6799"} Dec 09 09:50:41 crc kubenswrapper[4824]: I1209 09:50:41.275695 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"0d7422fc-b0b4-4118-a0de-980eaf4f4fad","Type":"ContainerStarted","Data":"ae3e5f27781c40b3bcdb38e1c8992b91ace8a97a34c07c7f1c81ac59412c3080"} Dec 09 09:50:41 crc kubenswrapper[4824]: I1209 09:50:41.278486 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-5tn52" event={"ID":"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219","Type":"ContainerStarted","Data":"e4064003824ca2b6fafa3a11661bcc5b9fd1781fdca0cb0e3f55fc1670499272"} Dec 09 09:50:41 crc kubenswrapper[4824]: I1209 09:50:41.278527 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-5tn52" event={"ID":"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219","Type":"ContainerStarted","Data":"68db5980c43569be9dcafd56ea0d8e3cb237439d45c7cf5e38fc56c776457498"} Dec 09 09:50:41 crc kubenswrapper[4824]: I1209 09:50:41.278542 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-5tn52" event={"ID":"fcbce9ea-3d55-4e09-b5bc-d5bb8c472219","Type":"ContainerStarted","Data":"1b894aefaaebc57e7d0b8f268113fd7f180d94ccee69c8e6ba8e138edf1e899a"} Dec 09 09:50:41 crc kubenswrapper[4824]: I1209 09:50:41.293942 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=8.293895536 podStartE2EDuration="8.293895536s" podCreationTimestamp="2025-12-09 09:50:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:50:41.292341074 +0000 UTC m=+197.626845741" watchObservedRunningTime="2025-12-09 09:50:41.293895536 +0000 UTC m=+197.628400203" Dec 09 09:50:41 crc kubenswrapper[4824]: I1209 09:50:41.334466 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=14.334447172 podStartE2EDuration="14.334447172s" podCreationTimestamp="2025-12-09 09:50:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:50:41.332085235 +0000 UTC m=+197.666589902" watchObservedRunningTime="2025-12-09 09:50:41.334447172 +0000 UTC m=+197.668951859" Dec 09 09:50:41 crc kubenswrapper[4824]: I1209 09:50:41.338761 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-5tn52" podStartSLOduration=177.338745383 podStartE2EDuration="2m57.338745383s" podCreationTimestamp="2025-12-09 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:50:41.316969661 +0000 UTC m=+197.651474338" watchObservedRunningTime="2025-12-09 09:50:41.338745383 +0000 UTC m=+197.673250060" Dec 09 09:50:42 crc kubenswrapper[4824]: I1209 09:50:42.286807 4824 generic.go:334] "Generic (PLEG): container finished" podID="0d7422fc-b0b4-4118-a0de-980eaf4f4fad" containerID="70c9ba842d3301c8024d030476c0b5efc53c125d19cebc3cb2dc5716d21f6799" exitCode=0 Dec 09 09:50:42 crc kubenswrapper[4824]: I1209 09:50:42.287229 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"0d7422fc-b0b4-4118-a0de-980eaf4f4fad","Type":"ContainerDied","Data":"70c9ba842d3301c8024d030476c0b5efc53c125d19cebc3cb2dc5716d21f6799"} Dec 09 09:50:43 crc kubenswrapper[4824]: I1209 09:50:43.555335 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 09:50:43 crc kubenswrapper[4824]: I1209 09:50:43.587968 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0d7422fc-b0b4-4118-a0de-980eaf4f4fad-kubelet-dir\") pod \"0d7422fc-b0b4-4118-a0de-980eaf4f4fad\" (UID: \"0d7422fc-b0b4-4118-a0de-980eaf4f4fad\") " Dec 09 09:50:43 crc kubenswrapper[4824]: I1209 09:50:43.588205 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d7422fc-b0b4-4118-a0de-980eaf4f4fad-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0d7422fc-b0b4-4118-a0de-980eaf4f4fad" (UID: "0d7422fc-b0b4-4118-a0de-980eaf4f4fad"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:50:43 crc kubenswrapper[4824]: I1209 09:50:43.689939 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0d7422fc-b0b4-4118-a0de-980eaf4f4fad-kube-api-access\") pod \"0d7422fc-b0b4-4118-a0de-980eaf4f4fad\" (UID: \"0d7422fc-b0b4-4118-a0de-980eaf4f4fad\") " Dec 09 09:50:43 crc kubenswrapper[4824]: I1209 09:50:43.690283 4824 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0d7422fc-b0b4-4118-a0de-980eaf4f4fad-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 09 09:50:43 crc kubenswrapper[4824]: I1209 09:50:43.705729 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d7422fc-b0b4-4118-a0de-980eaf4f4fad-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0d7422fc-b0b4-4118-a0de-980eaf4f4fad" (UID: "0d7422fc-b0b4-4118-a0de-980eaf4f4fad"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:50:43 crc kubenswrapper[4824]: I1209 09:50:43.792337 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0d7422fc-b0b4-4118-a0de-980eaf4f4fad-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 09:50:44 crc kubenswrapper[4824]: I1209 09:50:44.302406 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 09:50:44 crc kubenswrapper[4824]: I1209 09:50:44.304015 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"0d7422fc-b0b4-4118-a0de-980eaf4f4fad","Type":"ContainerDied","Data":"ae3e5f27781c40b3bcdb38e1c8992b91ace8a97a34c07c7f1c81ac59412c3080"} Dec 09 09:50:44 crc kubenswrapper[4824]: I1209 09:50:44.304093 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae3e5f27781c40b3bcdb38e1c8992b91ace8a97a34c07c7f1c81ac59412c3080" Dec 09 09:50:53 crc kubenswrapper[4824]: I1209 09:50:53.364517 4824 generic.go:334] "Generic (PLEG): container finished" podID="ad53b3fe-0e9e-4d56-9ca0-212dde3ac669" containerID="f69c0128e93ba119db30d9783dccc2f64676b2589a6880802057a6195fa2f957" exitCode=0 Dec 09 09:50:53 crc kubenswrapper[4824]: I1209 09:50:53.364627 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xg5qv" event={"ID":"ad53b3fe-0e9e-4d56-9ca0-212dde3ac669","Type":"ContainerDied","Data":"f69c0128e93ba119db30d9783dccc2f64676b2589a6880802057a6195fa2f957"} Dec 09 09:50:54 crc kubenswrapper[4824]: I1209 09:50:54.374330 4824 generic.go:334] "Generic (PLEG): container finished" podID="58792ed4-e360-4827-bf8c-508146205b78" containerID="3bd296d6db16386b94a4d82a7b737b3623c71538b2dbc27f60b5e97b616fca4a" exitCode=0 Dec 09 09:50:54 crc kubenswrapper[4824]: I1209 09:50:54.374448 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7pg7" event={"ID":"58792ed4-e360-4827-bf8c-508146205b78","Type":"ContainerDied","Data":"3bd296d6db16386b94a4d82a7b737b3623c71538b2dbc27f60b5e97b616fca4a"} Dec 09 09:50:54 crc kubenswrapper[4824]: I1209 09:50:54.377579 4824 generic.go:334] "Generic (PLEG): container finished" podID="e43afd8e-2c79-454d-9c95-a0f65f13dc37" containerID="3cf30e5c5b431173750f27940721bff0f91d8c47394ae13b0a23c2d06b82c22b" exitCode=0 Dec 09 09:50:54 crc kubenswrapper[4824]: I1209 09:50:54.377731 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hrhmc" event={"ID":"e43afd8e-2c79-454d-9c95-a0f65f13dc37","Type":"ContainerDied","Data":"3cf30e5c5b431173750f27940721bff0f91d8c47394ae13b0a23c2d06b82c22b"} Dec 09 09:50:54 crc kubenswrapper[4824]: I1209 09:50:54.380957 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4swl5" event={"ID":"7a6c3ca8-ba36-4cbc-9143-4f83a611144a","Type":"ContainerStarted","Data":"3d5718e25526c4ea67a3a8207fce24b4ef0abbc4b26b8a11d8b0e8f74955a4fc"} Dec 09 09:50:54 crc kubenswrapper[4824]: I1209 09:50:54.385828 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xg5qv" event={"ID":"ad53b3fe-0e9e-4d56-9ca0-212dde3ac669","Type":"ContainerStarted","Data":"5432c5f51e19a54172d29fded8f7a1a9c7251a0093fa48f470706bec986850a4"} Dec 09 09:50:54 crc kubenswrapper[4824]: I1209 09:50:54.389882 4824 generic.go:334] "Generic (PLEG): container finished" podID="e799b94c-33db-4de6-afd4-a8ffc4e9b069" containerID="f8579e4d3233a7272513eb992802b40a7d632811c573387c5167ef301b317979" exitCode=0 Dec 09 09:50:54 crc kubenswrapper[4824]: I1209 09:50:54.389978 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zmxrv" event={"ID":"e799b94c-33db-4de6-afd4-a8ffc4e9b069","Type":"ContainerDied","Data":"f8579e4d3233a7272513eb992802b40a7d632811c573387c5167ef301b317979"} Dec 09 09:50:54 crc kubenswrapper[4824]: I1209 09:50:54.392357 4824 generic.go:334] "Generic (PLEG): container finished" podID="db041858-ed93-4df9-a5d2-f79b7f9c3275" containerID="958189bac093637292ef3b226b1e67ef6dc51a7d5dd646cbb02881822d51ecfb" exitCode=0 Dec 09 09:50:54 crc kubenswrapper[4824]: I1209 09:50:54.392382 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9pdnx" event={"ID":"db041858-ed93-4df9-a5d2-f79b7f9c3275","Type":"ContainerDied","Data":"958189bac093637292ef3b226b1e67ef6dc51a7d5dd646cbb02881822d51ecfb"} Dec 09 09:50:54 crc kubenswrapper[4824]: I1209 09:50:54.523131 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xg5qv" podStartSLOduration=2.76533243 podStartE2EDuration="1m4.523073194s" podCreationTimestamp="2025-12-09 09:49:50 +0000 UTC" firstStartedPulling="2025-12-09 09:49:52.067474284 +0000 UTC m=+148.401978951" lastFinishedPulling="2025-12-09 09:50:53.825215048 +0000 UTC m=+210.159719715" observedRunningTime="2025-12-09 09:50:54.518372349 +0000 UTC m=+210.852877016" watchObservedRunningTime="2025-12-09 09:50:54.523073194 +0000 UTC m=+210.857577861" Dec 09 09:50:55 crc kubenswrapper[4824]: I1209 09:50:55.404609 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hrhmc" event={"ID":"e43afd8e-2c79-454d-9c95-a0f65f13dc37","Type":"ContainerStarted","Data":"71f71f644742164c36b08126f9061d9c099aaccd3e0094b96af3cf7caa946280"} Dec 09 09:50:55 crc kubenswrapper[4824]: I1209 09:50:55.407599 4824 generic.go:334] "Generic (PLEG): container finished" podID="7a6c3ca8-ba36-4cbc-9143-4f83a611144a" containerID="3d5718e25526c4ea67a3a8207fce24b4ef0abbc4b26b8a11d8b0e8f74955a4fc" exitCode=0 Dec 09 09:50:55 crc kubenswrapper[4824]: I1209 09:50:55.407721 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4swl5" event={"ID":"7a6c3ca8-ba36-4cbc-9143-4f83a611144a","Type":"ContainerDied","Data":"3d5718e25526c4ea67a3a8207fce24b4ef0abbc4b26b8a11d8b0e8f74955a4fc"} Dec 09 09:50:55 crc kubenswrapper[4824]: I1209 09:50:55.413470 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zmxrv" event={"ID":"e799b94c-33db-4de6-afd4-a8ffc4e9b069","Type":"ContainerStarted","Data":"7ffb8642d1d49d8e285914bb967a315745b78b0c10c9a28c4ee530ea14536831"} Dec 09 09:50:55 crc kubenswrapper[4824]: I1209 09:50:55.416887 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9pdnx" event={"ID":"db041858-ed93-4df9-a5d2-f79b7f9c3275","Type":"ContainerStarted","Data":"ee4c0468d0910dde9ab8bb7d7aa26531c2e01450ec95e3bfdff6a36b186654ea"} Dec 09 09:50:55 crc kubenswrapper[4824]: I1209 09:50:55.419747 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7pg7" event={"ID":"58792ed4-e360-4827-bf8c-508146205b78","Type":"ContainerStarted","Data":"6cffc534d451ede9615b562a1d9b2686276ce832fde7af2884e85d48b0e45c85"} Dec 09 09:50:55 crc kubenswrapper[4824]: I1209 09:50:55.438728 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hrhmc" podStartSLOduration=2.755529744 podStartE2EDuration="1m3.438698394s" podCreationTimestamp="2025-12-09 09:49:52 +0000 UTC" firstStartedPulling="2025-12-09 09:49:54.232685185 +0000 UTC m=+150.567189852" lastFinishedPulling="2025-12-09 09:50:54.915853835 +0000 UTC m=+211.250358502" observedRunningTime="2025-12-09 09:50:55.430060492 +0000 UTC m=+211.764565169" watchObservedRunningTime="2025-12-09 09:50:55.438698394 +0000 UTC m=+211.773203061" Dec 09 09:50:55 crc kubenswrapper[4824]: I1209 09:50:55.468944 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9pdnx" podStartSLOduration=2.410627133 podStartE2EDuration="1m6.468895983s" podCreationTimestamp="2025-12-09 09:49:49 +0000 UTC" firstStartedPulling="2025-12-09 09:49:51.061503798 +0000 UTC m=+147.396008465" lastFinishedPulling="2025-12-09 09:50:55.119772648 +0000 UTC m=+211.454277315" observedRunningTime="2025-12-09 09:50:55.462181793 +0000 UTC m=+211.796686470" watchObservedRunningTime="2025-12-09 09:50:55.468895983 +0000 UTC m=+211.803400640" Dec 09 09:50:55 crc kubenswrapper[4824]: I1209 09:50:55.504832 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-j7pg7" podStartSLOduration=2.623815572 podStartE2EDuration="1m4.504801168s" podCreationTimestamp="2025-12-09 09:49:51 +0000 UTC" firstStartedPulling="2025-12-09 09:49:53.097850661 +0000 UTC m=+149.432355318" lastFinishedPulling="2025-12-09 09:50:54.978836247 +0000 UTC m=+211.313340914" observedRunningTime="2025-12-09 09:50:55.502619936 +0000 UTC m=+211.837124603" watchObservedRunningTime="2025-12-09 09:50:55.504801168 +0000 UTC m=+211.839305855" Dec 09 09:50:55 crc kubenswrapper[4824]: I1209 09:50:55.548838 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zmxrv" podStartSLOduration=3.527681763 podStartE2EDuration="1m6.548809537s" podCreationTimestamp="2025-12-09 09:49:49 +0000 UTC" firstStartedPulling="2025-12-09 09:49:52.069069117 +0000 UTC m=+148.403573774" lastFinishedPulling="2025-12-09 09:50:55.090196881 +0000 UTC m=+211.424701548" observedRunningTime="2025-12-09 09:50:55.543873696 +0000 UTC m=+211.878378383" watchObservedRunningTime="2025-12-09 09:50:55.548809537 +0000 UTC m=+211.883314214" Dec 09 09:50:56 crc kubenswrapper[4824]: I1209 09:50:56.457820 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4swl5" event={"ID":"7a6c3ca8-ba36-4cbc-9143-4f83a611144a","Type":"ContainerStarted","Data":"2f1e63db81523f4b1630d3e37f6bed83b3a6a8ebcca3949a075d1fda98676a9f"} Dec 09 09:50:56 crc kubenswrapper[4824]: I1209 09:50:56.463353 4824 generic.go:334] "Generic (PLEG): container finished" podID="741767ab-ed9f-4b68-83bb-41aa25497741" containerID="082ab047d0517819a6cf6b09c30087eca66647f8a9d48e6777e39d727ced3678" exitCode=0 Dec 09 09:50:56 crc kubenswrapper[4824]: I1209 09:50:56.463631 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g9g6w" event={"ID":"741767ab-ed9f-4b68-83bb-41aa25497741","Type":"ContainerDied","Data":"082ab047d0517819a6cf6b09c30087eca66647f8a9d48e6777e39d727ced3678"} Dec 09 09:50:56 crc kubenswrapper[4824]: I1209 09:50:56.468529 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qpbgg" event={"ID":"3715eab2-b516-4d39-a23f-664a70efe5b8","Type":"ContainerStarted","Data":"1393f881c6354f5700898a5abe61086d31f974db1a007d8e11b201dad1e5d3bd"} Dec 09 09:50:56 crc kubenswrapper[4824]: I1209 09:50:56.520616 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4swl5" podStartSLOduration=2.809976828 podStartE2EDuration="1m3.520569805s" podCreationTimestamp="2025-12-09 09:49:53 +0000 UTC" firstStartedPulling="2025-12-09 09:49:55.297686055 +0000 UTC m=+151.632190722" lastFinishedPulling="2025-12-09 09:50:56.008279032 +0000 UTC m=+212.342783699" observedRunningTime="2025-12-09 09:50:56.487501484 +0000 UTC m=+212.822006151" watchObservedRunningTime="2025-12-09 09:50:56.520569805 +0000 UTC m=+212.855074472" Dec 09 09:50:57 crc kubenswrapper[4824]: I1209 09:50:57.479308 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g9g6w" event={"ID":"741767ab-ed9f-4b68-83bb-41aa25497741","Type":"ContainerStarted","Data":"5e4989257e5025288275554e3c16052be2bfdc1300fb33d2662dbd9046884c97"} Dec 09 09:50:57 crc kubenswrapper[4824]: I1209 09:50:57.501307 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-g9g6w" podStartSLOduration=3.590754428 podStartE2EDuration="1m8.501286537s" podCreationTimestamp="2025-12-09 09:49:49 +0000 UTC" firstStartedPulling="2025-12-09 09:49:52.07070114 +0000 UTC m=+148.405205807" lastFinishedPulling="2025-12-09 09:50:56.981233249 +0000 UTC m=+213.315737916" observedRunningTime="2025-12-09 09:50:57.499133386 +0000 UTC m=+213.833638073" watchObservedRunningTime="2025-12-09 09:50:57.501286537 +0000 UTC m=+213.835791204" Dec 09 09:50:58 crc kubenswrapper[4824]: I1209 09:50:58.489402 4824 generic.go:334] "Generic (PLEG): container finished" podID="3715eab2-b516-4d39-a23f-664a70efe5b8" containerID="1393f881c6354f5700898a5abe61086d31f974db1a007d8e11b201dad1e5d3bd" exitCode=0 Dec 09 09:50:58 crc kubenswrapper[4824]: I1209 09:50:58.489502 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qpbgg" event={"ID":"3715eab2-b516-4d39-a23f-664a70efe5b8","Type":"ContainerDied","Data":"1393f881c6354f5700898a5abe61086d31f974db1a007d8e11b201dad1e5d3bd"} Dec 09 09:50:59 crc kubenswrapper[4824]: I1209 09:50:59.918994 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9pdnx" Dec 09 09:50:59 crc kubenswrapper[4824]: I1209 09:50:59.919429 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9pdnx" Dec 09 09:50:59 crc kubenswrapper[4824]: I1209 09:50:59.986063 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9pdnx" Dec 09 09:51:00 crc kubenswrapper[4824]: I1209 09:51:00.386483 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-g9g6w" Dec 09 09:51:00 crc kubenswrapper[4824]: I1209 09:51:00.386557 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-g9g6w" Dec 09 09:51:00 crc kubenswrapper[4824]: I1209 09:51:00.438970 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zmxrv" Dec 09 09:51:00 crc kubenswrapper[4824]: I1209 09:51:00.439035 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zmxrv" Dec 09 09:51:00 crc kubenswrapper[4824]: I1209 09:51:00.505243 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qpbgg" event={"ID":"3715eab2-b516-4d39-a23f-664a70efe5b8","Type":"ContainerStarted","Data":"2e92fd06942e02ae8b3e4a61127f817724523a7371da5f756c925e65a0db72a8"} Dec 09 09:51:00 crc kubenswrapper[4824]: I1209 09:51:00.557335 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zmxrv" Dec 09 09:51:00 crc kubenswrapper[4824]: I1209 09:51:00.590557 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qpbgg" podStartSLOduration=3.6652487320000002 podStartE2EDuration="1m8.590531973s" podCreationTimestamp="2025-12-09 09:49:52 +0000 UTC" firstStartedPulling="2025-12-09 09:49:54.196965107 +0000 UTC m=+150.531469774" lastFinishedPulling="2025-12-09 09:50:59.122248348 +0000 UTC m=+215.456753015" observedRunningTime="2025-12-09 09:51:00.564527033 +0000 UTC m=+216.899031700" watchObservedRunningTime="2025-12-09 09:51:00.590531973 +0000 UTC m=+216.925036640" Dec 09 09:51:00 crc kubenswrapper[4824]: I1209 09:51:00.614499 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xg5qv" Dec 09 09:51:00 crc kubenswrapper[4824]: I1209 09:51:00.614559 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xg5qv" Dec 09 09:51:00 crc kubenswrapper[4824]: I1209 09:51:00.672304 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-g9g6w" Dec 09 09:51:00 crc kubenswrapper[4824]: I1209 09:51:00.712409 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zmxrv" Dec 09 09:51:00 crc kubenswrapper[4824]: I1209 09:51:00.723812 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xg5qv" Dec 09 09:51:00 crc kubenswrapper[4824]: I1209 09:51:00.773006 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9pdnx" Dec 09 09:51:01 crc kubenswrapper[4824]: I1209 09:51:01.559227 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xg5qv" Dec 09 09:51:02 crc kubenswrapper[4824]: I1209 09:51:02.129579 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-j7pg7" Dec 09 09:51:02 crc kubenswrapper[4824]: I1209 09:51:02.130161 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-j7pg7" Dec 09 09:51:02 crc kubenswrapper[4824]: I1209 09:51:02.179000 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-j7pg7" Dec 09 09:51:02 crc kubenswrapper[4824]: I1209 09:51:02.558980 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-j7pg7" Dec 09 09:51:02 crc kubenswrapper[4824]: I1209 09:51:02.572438 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hrhmc" Dec 09 09:51:02 crc kubenswrapper[4824]: I1209 09:51:02.572630 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hrhmc" Dec 09 09:51:02 crc kubenswrapper[4824]: I1209 09:51:02.632103 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hrhmc" Dec 09 09:51:02 crc kubenswrapper[4824]: I1209 09:51:02.861476 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 09:51:02 crc kubenswrapper[4824]: I1209 09:51:02.861565 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 09:51:02 crc kubenswrapper[4824]: I1209 09:51:02.861630 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 09:51:02 crc kubenswrapper[4824]: I1209 09:51:02.862483 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671"} pod="openshift-machine-config-operator/machine-config-daemon-dth8x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 09:51:02 crc kubenswrapper[4824]: I1209 09:51:02.862615 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" containerID="cri-o://314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671" gracePeriod=600 Dec 09 09:51:03 crc kubenswrapper[4824]: I1209 09:51:03.137645 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qpbgg" Dec 09 09:51:03 crc kubenswrapper[4824]: I1209 09:51:03.138017 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qpbgg" Dec 09 09:51:03 crc kubenswrapper[4824]: I1209 09:51:03.362456 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xg5qv"] Dec 09 09:51:03 crc kubenswrapper[4824]: I1209 09:51:03.523652 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerID="314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671" exitCode=0 Dec 09 09:51:03 crc kubenswrapper[4824]: I1209 09:51:03.523738 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerDied","Data":"314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671"} Dec 09 09:51:03 crc kubenswrapper[4824]: I1209 09:51:03.524672 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xg5qv" podUID="ad53b3fe-0e9e-4d56-9ca0-212dde3ac669" containerName="registry-server" containerID="cri-o://5432c5f51e19a54172d29fded8f7a1a9c7251a0093fa48f470706bec986850a4" gracePeriod=2 Dec 09 09:51:03 crc kubenswrapper[4824]: I1209 09:51:03.541314 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4swl5" Dec 09 09:51:03 crc kubenswrapper[4824]: I1209 09:51:03.541378 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4swl5" Dec 09 09:51:03 crc kubenswrapper[4824]: I1209 09:51:03.581239 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hrhmc" Dec 09 09:51:03 crc kubenswrapper[4824]: I1209 09:51:03.591406 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4swl5" Dec 09 09:51:03 crc kubenswrapper[4824]: I1209 09:51:03.974549 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xg5qv" Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.015741 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad53b3fe-0e9e-4d56-9ca0-212dde3ac669-utilities\") pod \"ad53b3fe-0e9e-4d56-9ca0-212dde3ac669\" (UID: \"ad53b3fe-0e9e-4d56-9ca0-212dde3ac669\") " Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.015909 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qx6hv\" (UniqueName: \"kubernetes.io/projected/ad53b3fe-0e9e-4d56-9ca0-212dde3ac669-kube-api-access-qx6hv\") pod \"ad53b3fe-0e9e-4d56-9ca0-212dde3ac669\" (UID: \"ad53b3fe-0e9e-4d56-9ca0-212dde3ac669\") " Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.015967 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad53b3fe-0e9e-4d56-9ca0-212dde3ac669-catalog-content\") pod \"ad53b3fe-0e9e-4d56-9ca0-212dde3ac669\" (UID: \"ad53b3fe-0e9e-4d56-9ca0-212dde3ac669\") " Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.018839 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad53b3fe-0e9e-4d56-9ca0-212dde3ac669-utilities" (OuterVolumeSpecName: "utilities") pod "ad53b3fe-0e9e-4d56-9ca0-212dde3ac669" (UID: "ad53b3fe-0e9e-4d56-9ca0-212dde3ac669"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.026906 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad53b3fe-0e9e-4d56-9ca0-212dde3ac669-kube-api-access-qx6hv" (OuterVolumeSpecName: "kube-api-access-qx6hv") pod "ad53b3fe-0e9e-4d56-9ca0-212dde3ac669" (UID: "ad53b3fe-0e9e-4d56-9ca0-212dde3ac669"). InnerVolumeSpecName "kube-api-access-qx6hv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.111253 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad53b3fe-0e9e-4d56-9ca0-212dde3ac669-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad53b3fe-0e9e-4d56-9ca0-212dde3ac669" (UID: "ad53b3fe-0e9e-4d56-9ca0-212dde3ac669"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.118399 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad53b3fe-0e9e-4d56-9ca0-212dde3ac669-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.118456 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qx6hv\" (UniqueName: \"kubernetes.io/projected/ad53b3fe-0e9e-4d56-9ca0-212dde3ac669-kube-api-access-qx6hv\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.118471 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad53b3fe-0e9e-4d56-9ca0-212dde3ac669-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.192296 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qpbgg" podUID="3715eab2-b516-4d39-a23f-664a70efe5b8" containerName="registry-server" probeResult="failure" output=< Dec 09 09:51:04 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 09:51:04 crc kubenswrapper[4824]: > Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.533616 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerStarted","Data":"ea7720fbe70782cdf3ecb70ae54c936954803442a5c7d2c6bc34bd0e635bfbfe"} Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.535860 4824 generic.go:334] "Generic (PLEG): container finished" podID="ad53b3fe-0e9e-4d56-9ca0-212dde3ac669" containerID="5432c5f51e19a54172d29fded8f7a1a9c7251a0093fa48f470706bec986850a4" exitCode=0 Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.536479 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xg5qv" Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.537074 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xg5qv" event={"ID":"ad53b3fe-0e9e-4d56-9ca0-212dde3ac669","Type":"ContainerDied","Data":"5432c5f51e19a54172d29fded8f7a1a9c7251a0093fa48f470706bec986850a4"} Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.537151 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xg5qv" event={"ID":"ad53b3fe-0e9e-4d56-9ca0-212dde3ac669","Type":"ContainerDied","Data":"0b2834a7c7988d5f82185d739f6af31e95a8864c29673c0fba63dee5b8ffab79"} Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.537220 4824 scope.go:117] "RemoveContainer" containerID="5432c5f51e19a54172d29fded8f7a1a9c7251a0093fa48f470706bec986850a4" Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.571140 4824 scope.go:117] "RemoveContainer" containerID="f69c0128e93ba119db30d9783dccc2f64676b2589a6880802057a6195fa2f957" Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.606437 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4swl5" Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.614747 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xg5qv"] Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.627912 4824 scope.go:117] "RemoveContainer" containerID="6501aeec075b8a0c31f3c5dbc033f30ecdc3cba0674dde35b9bae12572492c80" Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.638355 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xg5qv"] Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.669484 4824 scope.go:117] "RemoveContainer" containerID="5432c5f51e19a54172d29fded8f7a1a9c7251a0093fa48f470706bec986850a4" Dec 09 09:51:04 crc kubenswrapper[4824]: E1209 09:51:04.670073 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5432c5f51e19a54172d29fded8f7a1a9c7251a0093fa48f470706bec986850a4\": container with ID starting with 5432c5f51e19a54172d29fded8f7a1a9c7251a0093fa48f470706bec986850a4 not found: ID does not exist" containerID="5432c5f51e19a54172d29fded8f7a1a9c7251a0093fa48f470706bec986850a4" Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.670200 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5432c5f51e19a54172d29fded8f7a1a9c7251a0093fa48f470706bec986850a4"} err="failed to get container status \"5432c5f51e19a54172d29fded8f7a1a9c7251a0093fa48f470706bec986850a4\": rpc error: code = NotFound desc = could not find container \"5432c5f51e19a54172d29fded8f7a1a9c7251a0093fa48f470706bec986850a4\": container with ID starting with 5432c5f51e19a54172d29fded8f7a1a9c7251a0093fa48f470706bec986850a4 not found: ID does not exist" Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.670323 4824 scope.go:117] "RemoveContainer" containerID="f69c0128e93ba119db30d9783dccc2f64676b2589a6880802057a6195fa2f957" Dec 09 09:51:04 crc kubenswrapper[4824]: E1209 09:51:04.671401 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f69c0128e93ba119db30d9783dccc2f64676b2589a6880802057a6195fa2f957\": container with ID starting with f69c0128e93ba119db30d9783dccc2f64676b2589a6880802057a6195fa2f957 not found: ID does not exist" containerID="f69c0128e93ba119db30d9783dccc2f64676b2589a6880802057a6195fa2f957" Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.671466 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f69c0128e93ba119db30d9783dccc2f64676b2589a6880802057a6195fa2f957"} err="failed to get container status \"f69c0128e93ba119db30d9783dccc2f64676b2589a6880802057a6195fa2f957\": rpc error: code = NotFound desc = could not find container \"f69c0128e93ba119db30d9783dccc2f64676b2589a6880802057a6195fa2f957\": container with ID starting with f69c0128e93ba119db30d9783dccc2f64676b2589a6880802057a6195fa2f957 not found: ID does not exist" Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.671521 4824 scope.go:117] "RemoveContainer" containerID="6501aeec075b8a0c31f3c5dbc033f30ecdc3cba0674dde35b9bae12572492c80" Dec 09 09:51:04 crc kubenswrapper[4824]: E1209 09:51:04.671878 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6501aeec075b8a0c31f3c5dbc033f30ecdc3cba0674dde35b9bae12572492c80\": container with ID starting with 6501aeec075b8a0c31f3c5dbc033f30ecdc3cba0674dde35b9bae12572492c80 not found: ID does not exist" containerID="6501aeec075b8a0c31f3c5dbc033f30ecdc3cba0674dde35b9bae12572492c80" Dec 09 09:51:04 crc kubenswrapper[4824]: I1209 09:51:04.672359 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6501aeec075b8a0c31f3c5dbc033f30ecdc3cba0674dde35b9bae12572492c80"} err="failed to get container status \"6501aeec075b8a0c31f3c5dbc033f30ecdc3cba0674dde35b9bae12572492c80\": rpc error: code = NotFound desc = could not find container \"6501aeec075b8a0c31f3c5dbc033f30ecdc3cba0674dde35b9bae12572492c80\": container with ID starting with 6501aeec075b8a0c31f3c5dbc033f30ecdc3cba0674dde35b9bae12572492c80 not found: ID does not exist" Dec 09 09:51:05 crc kubenswrapper[4824]: I1209 09:51:05.758709 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hrhmc"] Dec 09 09:51:05 crc kubenswrapper[4824]: I1209 09:51:05.918455 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad53b3fe-0e9e-4d56-9ca0-212dde3ac669" path="/var/lib/kubelet/pods/ad53b3fe-0e9e-4d56-9ca0-212dde3ac669/volumes" Dec 09 09:51:06 crc kubenswrapper[4824]: I1209 09:51:06.548283 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hrhmc" podUID="e43afd8e-2c79-454d-9c95-a0f65f13dc37" containerName="registry-server" containerID="cri-o://71f71f644742164c36b08126f9061d9c099aaccd3e0094b96af3cf7caa946280" gracePeriod=2 Dec 09 09:51:06 crc kubenswrapper[4824]: I1209 09:51:06.930161 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hrhmc" Dec 09 09:51:06 crc kubenswrapper[4824]: I1209 09:51:06.964207 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e43afd8e-2c79-454d-9c95-a0f65f13dc37-utilities\") pod \"e43afd8e-2c79-454d-9c95-a0f65f13dc37\" (UID: \"e43afd8e-2c79-454d-9c95-a0f65f13dc37\") " Dec 09 09:51:06 crc kubenswrapper[4824]: I1209 09:51:06.964297 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxsnt\" (UniqueName: \"kubernetes.io/projected/e43afd8e-2c79-454d-9c95-a0f65f13dc37-kube-api-access-bxsnt\") pod \"e43afd8e-2c79-454d-9c95-a0f65f13dc37\" (UID: \"e43afd8e-2c79-454d-9c95-a0f65f13dc37\") " Dec 09 09:51:06 crc kubenswrapper[4824]: I1209 09:51:06.964343 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e43afd8e-2c79-454d-9c95-a0f65f13dc37-catalog-content\") pod \"e43afd8e-2c79-454d-9c95-a0f65f13dc37\" (UID: \"e43afd8e-2c79-454d-9c95-a0f65f13dc37\") " Dec 09 09:51:06 crc kubenswrapper[4824]: I1209 09:51:06.965122 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e43afd8e-2c79-454d-9c95-a0f65f13dc37-utilities" (OuterVolumeSpecName: "utilities") pod "e43afd8e-2c79-454d-9c95-a0f65f13dc37" (UID: "e43afd8e-2c79-454d-9c95-a0f65f13dc37"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:51:06 crc kubenswrapper[4824]: I1209 09:51:06.974977 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e43afd8e-2c79-454d-9c95-a0f65f13dc37-kube-api-access-bxsnt" (OuterVolumeSpecName: "kube-api-access-bxsnt") pod "e43afd8e-2c79-454d-9c95-a0f65f13dc37" (UID: "e43afd8e-2c79-454d-9c95-a0f65f13dc37"). InnerVolumeSpecName "kube-api-access-bxsnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:51:06 crc kubenswrapper[4824]: I1209 09:51:06.988149 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e43afd8e-2c79-454d-9c95-a0f65f13dc37-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e43afd8e-2c79-454d-9c95-a0f65f13dc37" (UID: "e43afd8e-2c79-454d-9c95-a0f65f13dc37"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:51:07 crc kubenswrapper[4824]: I1209 09:51:07.065937 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxsnt\" (UniqueName: \"kubernetes.io/projected/e43afd8e-2c79-454d-9c95-a0f65f13dc37-kube-api-access-bxsnt\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:07 crc kubenswrapper[4824]: I1209 09:51:07.065980 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e43afd8e-2c79-454d-9c95-a0f65f13dc37-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:07 crc kubenswrapper[4824]: I1209 09:51:07.065992 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e43afd8e-2c79-454d-9c95-a0f65f13dc37-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:07 crc kubenswrapper[4824]: I1209 09:51:07.558474 4824 generic.go:334] "Generic (PLEG): container finished" podID="e43afd8e-2c79-454d-9c95-a0f65f13dc37" containerID="71f71f644742164c36b08126f9061d9c099aaccd3e0094b96af3cf7caa946280" exitCode=0 Dec 09 09:51:07 crc kubenswrapper[4824]: I1209 09:51:07.558585 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hrhmc" Dec 09 09:51:07 crc kubenswrapper[4824]: I1209 09:51:07.558580 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hrhmc" event={"ID":"e43afd8e-2c79-454d-9c95-a0f65f13dc37","Type":"ContainerDied","Data":"71f71f644742164c36b08126f9061d9c099aaccd3e0094b96af3cf7caa946280"} Dec 09 09:51:07 crc kubenswrapper[4824]: I1209 09:51:07.559205 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hrhmc" event={"ID":"e43afd8e-2c79-454d-9c95-a0f65f13dc37","Type":"ContainerDied","Data":"51928036c9ebedbb7d909a164d67cd07f5c9ec1bf4a78371bcf4ee1da4d87999"} Dec 09 09:51:07 crc kubenswrapper[4824]: I1209 09:51:07.559237 4824 scope.go:117] "RemoveContainer" containerID="71f71f644742164c36b08126f9061d9c099aaccd3e0094b96af3cf7caa946280" Dec 09 09:51:07 crc kubenswrapper[4824]: I1209 09:51:07.594411 4824 scope.go:117] "RemoveContainer" containerID="3cf30e5c5b431173750f27940721bff0f91d8c47394ae13b0a23c2d06b82c22b" Dec 09 09:51:07 crc kubenswrapper[4824]: I1209 09:51:07.595798 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hrhmc"] Dec 09 09:51:07 crc kubenswrapper[4824]: I1209 09:51:07.601009 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hrhmc"] Dec 09 09:51:07 crc kubenswrapper[4824]: I1209 09:51:07.613641 4824 scope.go:117] "RemoveContainer" containerID="352c1aeab5048b8931929b28166d07fabb6fa498f223e192d9b3ffcbeb23603a" Dec 09 09:51:07 crc kubenswrapper[4824]: I1209 09:51:07.638320 4824 scope.go:117] "RemoveContainer" containerID="71f71f644742164c36b08126f9061d9c099aaccd3e0094b96af3cf7caa946280" Dec 09 09:51:07 crc kubenswrapper[4824]: E1209 09:51:07.639153 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71f71f644742164c36b08126f9061d9c099aaccd3e0094b96af3cf7caa946280\": container with ID starting with 71f71f644742164c36b08126f9061d9c099aaccd3e0094b96af3cf7caa946280 not found: ID does not exist" containerID="71f71f644742164c36b08126f9061d9c099aaccd3e0094b96af3cf7caa946280" Dec 09 09:51:07 crc kubenswrapper[4824]: I1209 09:51:07.639199 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71f71f644742164c36b08126f9061d9c099aaccd3e0094b96af3cf7caa946280"} err="failed to get container status \"71f71f644742164c36b08126f9061d9c099aaccd3e0094b96af3cf7caa946280\": rpc error: code = NotFound desc = could not find container \"71f71f644742164c36b08126f9061d9c099aaccd3e0094b96af3cf7caa946280\": container with ID starting with 71f71f644742164c36b08126f9061d9c099aaccd3e0094b96af3cf7caa946280 not found: ID does not exist" Dec 09 09:51:07 crc kubenswrapper[4824]: I1209 09:51:07.639258 4824 scope.go:117] "RemoveContainer" containerID="3cf30e5c5b431173750f27940721bff0f91d8c47394ae13b0a23c2d06b82c22b" Dec 09 09:51:07 crc kubenswrapper[4824]: E1209 09:51:07.639567 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cf30e5c5b431173750f27940721bff0f91d8c47394ae13b0a23c2d06b82c22b\": container with ID starting with 3cf30e5c5b431173750f27940721bff0f91d8c47394ae13b0a23c2d06b82c22b not found: ID does not exist" containerID="3cf30e5c5b431173750f27940721bff0f91d8c47394ae13b0a23c2d06b82c22b" Dec 09 09:51:07 crc kubenswrapper[4824]: I1209 09:51:07.639606 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cf30e5c5b431173750f27940721bff0f91d8c47394ae13b0a23c2d06b82c22b"} err="failed to get container status \"3cf30e5c5b431173750f27940721bff0f91d8c47394ae13b0a23c2d06b82c22b\": rpc error: code = NotFound desc = could not find container \"3cf30e5c5b431173750f27940721bff0f91d8c47394ae13b0a23c2d06b82c22b\": container with ID starting with 3cf30e5c5b431173750f27940721bff0f91d8c47394ae13b0a23c2d06b82c22b not found: ID does not exist" Dec 09 09:51:07 crc kubenswrapper[4824]: I1209 09:51:07.639634 4824 scope.go:117] "RemoveContainer" containerID="352c1aeab5048b8931929b28166d07fabb6fa498f223e192d9b3ffcbeb23603a" Dec 09 09:51:07 crc kubenswrapper[4824]: E1209 09:51:07.640202 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"352c1aeab5048b8931929b28166d07fabb6fa498f223e192d9b3ffcbeb23603a\": container with ID starting with 352c1aeab5048b8931929b28166d07fabb6fa498f223e192d9b3ffcbeb23603a not found: ID does not exist" containerID="352c1aeab5048b8931929b28166d07fabb6fa498f223e192d9b3ffcbeb23603a" Dec 09 09:51:07 crc kubenswrapper[4824]: I1209 09:51:07.640228 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"352c1aeab5048b8931929b28166d07fabb6fa498f223e192d9b3ffcbeb23603a"} err="failed to get container status \"352c1aeab5048b8931929b28166d07fabb6fa498f223e192d9b3ffcbeb23603a\": rpc error: code = NotFound desc = could not find container \"352c1aeab5048b8931929b28166d07fabb6fa498f223e192d9b3ffcbeb23603a\": container with ID starting with 352c1aeab5048b8931929b28166d07fabb6fa498f223e192d9b3ffcbeb23603a not found: ID does not exist" Dec 09 09:51:07 crc kubenswrapper[4824]: I1209 09:51:07.918271 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e43afd8e-2c79-454d-9c95-a0f65f13dc37" path="/var/lib/kubelet/pods/e43afd8e-2c79-454d-9c95-a0f65f13dc37/volumes" Dec 09 09:51:08 crc kubenswrapper[4824]: I1209 09:51:08.160446 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4swl5"] Dec 09 09:51:08 crc kubenswrapper[4824]: I1209 09:51:08.160813 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4swl5" podUID="7a6c3ca8-ba36-4cbc-9143-4f83a611144a" containerName="registry-server" containerID="cri-o://2f1e63db81523f4b1630d3e37f6bed83b3a6a8ebcca3949a075d1fda98676a9f" gracePeriod=2 Dec 09 09:51:08 crc kubenswrapper[4824]: I1209 09:51:08.865582 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4swl5" Dec 09 09:51:08 crc kubenswrapper[4824]: I1209 09:51:08.894703 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6c3ca8-ba36-4cbc-9143-4f83a611144a-utilities\") pod \"7a6c3ca8-ba36-4cbc-9143-4f83a611144a\" (UID: \"7a6c3ca8-ba36-4cbc-9143-4f83a611144a\") " Dec 09 09:51:08 crc kubenswrapper[4824]: I1209 09:51:08.894796 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6c3ca8-ba36-4cbc-9143-4f83a611144a-catalog-content\") pod \"7a6c3ca8-ba36-4cbc-9143-4f83a611144a\" (UID: \"7a6c3ca8-ba36-4cbc-9143-4f83a611144a\") " Dec 09 09:51:08 crc kubenswrapper[4824]: I1209 09:51:08.894902 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nj9rd\" (UniqueName: \"kubernetes.io/projected/7a6c3ca8-ba36-4cbc-9143-4f83a611144a-kube-api-access-nj9rd\") pod \"7a6c3ca8-ba36-4cbc-9143-4f83a611144a\" (UID: \"7a6c3ca8-ba36-4cbc-9143-4f83a611144a\") " Dec 09 09:51:08 crc kubenswrapper[4824]: I1209 09:51:08.895916 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a6c3ca8-ba36-4cbc-9143-4f83a611144a-utilities" (OuterVolumeSpecName: "utilities") pod "7a6c3ca8-ba36-4cbc-9143-4f83a611144a" (UID: "7a6c3ca8-ba36-4cbc-9143-4f83a611144a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:51:08 crc kubenswrapper[4824]: I1209 09:51:08.905771 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a6c3ca8-ba36-4cbc-9143-4f83a611144a-kube-api-access-nj9rd" (OuterVolumeSpecName: "kube-api-access-nj9rd") pod "7a6c3ca8-ba36-4cbc-9143-4f83a611144a" (UID: "7a6c3ca8-ba36-4cbc-9143-4f83a611144a"). InnerVolumeSpecName "kube-api-access-nj9rd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:51:08 crc kubenswrapper[4824]: I1209 09:51:08.996993 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nj9rd\" (UniqueName: \"kubernetes.io/projected/7a6c3ca8-ba36-4cbc-9143-4f83a611144a-kube-api-access-nj9rd\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:08 crc kubenswrapper[4824]: I1209 09:51:08.997048 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6c3ca8-ba36-4cbc-9143-4f83a611144a-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:09 crc kubenswrapper[4824]: I1209 09:51:09.040914 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a6c3ca8-ba36-4cbc-9143-4f83a611144a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a6c3ca8-ba36-4cbc-9143-4f83a611144a" (UID: "7a6c3ca8-ba36-4cbc-9143-4f83a611144a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:51:09 crc kubenswrapper[4824]: I1209 09:51:09.098413 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6c3ca8-ba36-4cbc-9143-4f83a611144a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:09 crc kubenswrapper[4824]: I1209 09:51:09.575829 4824 generic.go:334] "Generic (PLEG): container finished" podID="7a6c3ca8-ba36-4cbc-9143-4f83a611144a" containerID="2f1e63db81523f4b1630d3e37f6bed83b3a6a8ebcca3949a075d1fda98676a9f" exitCode=0 Dec 09 09:51:09 crc kubenswrapper[4824]: I1209 09:51:09.575924 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4swl5" Dec 09 09:51:09 crc kubenswrapper[4824]: I1209 09:51:09.575923 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4swl5" event={"ID":"7a6c3ca8-ba36-4cbc-9143-4f83a611144a","Type":"ContainerDied","Data":"2f1e63db81523f4b1630d3e37f6bed83b3a6a8ebcca3949a075d1fda98676a9f"} Dec 09 09:51:09 crc kubenswrapper[4824]: I1209 09:51:09.576456 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4swl5" event={"ID":"7a6c3ca8-ba36-4cbc-9143-4f83a611144a","Type":"ContainerDied","Data":"b247c8841ba6392effdc3604929acb2447df14980fb04bc3c8e77f03a3062f2f"} Dec 09 09:51:09 crc kubenswrapper[4824]: I1209 09:51:09.576481 4824 scope.go:117] "RemoveContainer" containerID="2f1e63db81523f4b1630d3e37f6bed83b3a6a8ebcca3949a075d1fda98676a9f" Dec 09 09:51:09 crc kubenswrapper[4824]: I1209 09:51:09.599604 4824 scope.go:117] "RemoveContainer" containerID="3d5718e25526c4ea67a3a8207fce24b4ef0abbc4b26b8a11d8b0e8f74955a4fc" Dec 09 09:51:09 crc kubenswrapper[4824]: I1209 09:51:09.614207 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4swl5"] Dec 09 09:51:09 crc kubenswrapper[4824]: I1209 09:51:09.616137 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4swl5"] Dec 09 09:51:09 crc kubenswrapper[4824]: I1209 09:51:09.634426 4824 scope.go:117] "RemoveContainer" containerID="fb95a63af9504846fdcf803d4dd2675bdbb03544b6030a50c4a146cf65b5448a" Dec 09 09:51:09 crc kubenswrapper[4824]: I1209 09:51:09.652042 4824 scope.go:117] "RemoveContainer" containerID="2f1e63db81523f4b1630d3e37f6bed83b3a6a8ebcca3949a075d1fda98676a9f" Dec 09 09:51:09 crc kubenswrapper[4824]: E1209 09:51:09.652704 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f1e63db81523f4b1630d3e37f6bed83b3a6a8ebcca3949a075d1fda98676a9f\": container with ID starting with 2f1e63db81523f4b1630d3e37f6bed83b3a6a8ebcca3949a075d1fda98676a9f not found: ID does not exist" containerID="2f1e63db81523f4b1630d3e37f6bed83b3a6a8ebcca3949a075d1fda98676a9f" Dec 09 09:51:09 crc kubenswrapper[4824]: I1209 09:51:09.652845 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f1e63db81523f4b1630d3e37f6bed83b3a6a8ebcca3949a075d1fda98676a9f"} err="failed to get container status \"2f1e63db81523f4b1630d3e37f6bed83b3a6a8ebcca3949a075d1fda98676a9f\": rpc error: code = NotFound desc = could not find container \"2f1e63db81523f4b1630d3e37f6bed83b3a6a8ebcca3949a075d1fda98676a9f\": container with ID starting with 2f1e63db81523f4b1630d3e37f6bed83b3a6a8ebcca3949a075d1fda98676a9f not found: ID does not exist" Dec 09 09:51:09 crc kubenswrapper[4824]: I1209 09:51:09.652973 4824 scope.go:117] "RemoveContainer" containerID="3d5718e25526c4ea67a3a8207fce24b4ef0abbc4b26b8a11d8b0e8f74955a4fc" Dec 09 09:51:09 crc kubenswrapper[4824]: E1209 09:51:09.653433 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d5718e25526c4ea67a3a8207fce24b4ef0abbc4b26b8a11d8b0e8f74955a4fc\": container with ID starting with 3d5718e25526c4ea67a3a8207fce24b4ef0abbc4b26b8a11d8b0e8f74955a4fc not found: ID does not exist" containerID="3d5718e25526c4ea67a3a8207fce24b4ef0abbc4b26b8a11d8b0e8f74955a4fc" Dec 09 09:51:09 crc kubenswrapper[4824]: I1209 09:51:09.653528 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d5718e25526c4ea67a3a8207fce24b4ef0abbc4b26b8a11d8b0e8f74955a4fc"} err="failed to get container status \"3d5718e25526c4ea67a3a8207fce24b4ef0abbc4b26b8a11d8b0e8f74955a4fc\": rpc error: code = NotFound desc = could not find container \"3d5718e25526c4ea67a3a8207fce24b4ef0abbc4b26b8a11d8b0e8f74955a4fc\": container with ID starting with 3d5718e25526c4ea67a3a8207fce24b4ef0abbc4b26b8a11d8b0e8f74955a4fc not found: ID does not exist" Dec 09 09:51:09 crc kubenswrapper[4824]: I1209 09:51:09.653599 4824 scope.go:117] "RemoveContainer" containerID="fb95a63af9504846fdcf803d4dd2675bdbb03544b6030a50c4a146cf65b5448a" Dec 09 09:51:09 crc kubenswrapper[4824]: E1209 09:51:09.653976 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb95a63af9504846fdcf803d4dd2675bdbb03544b6030a50c4a146cf65b5448a\": container with ID starting with fb95a63af9504846fdcf803d4dd2675bdbb03544b6030a50c4a146cf65b5448a not found: ID does not exist" containerID="fb95a63af9504846fdcf803d4dd2675bdbb03544b6030a50c4a146cf65b5448a" Dec 09 09:51:09 crc kubenswrapper[4824]: I1209 09:51:09.654061 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb95a63af9504846fdcf803d4dd2675bdbb03544b6030a50c4a146cf65b5448a"} err="failed to get container status \"fb95a63af9504846fdcf803d4dd2675bdbb03544b6030a50c4a146cf65b5448a\": rpc error: code = NotFound desc = could not find container \"fb95a63af9504846fdcf803d4dd2675bdbb03544b6030a50c4a146cf65b5448a\": container with ID starting with fb95a63af9504846fdcf803d4dd2675bdbb03544b6030a50c4a146cf65b5448a not found: ID does not exist" Dec 09 09:51:09 crc kubenswrapper[4824]: I1209 09:51:09.918865 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a6c3ca8-ba36-4cbc-9143-4f83a611144a" path="/var/lib/kubelet/pods/7a6c3ca8-ba36-4cbc-9143-4f83a611144a/volumes" Dec 09 09:51:09 crc kubenswrapper[4824]: I1209 09:51:09.959310 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-l5ld8"] Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.190754 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9pdnx"] Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.191421 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9pdnx" podUID="db041858-ed93-4df9-a5d2-f79b7f9c3275" containerName="registry-server" containerID="cri-o://ee4c0468d0910dde9ab8bb7d7aa26531c2e01450ec95e3bfdff6a36b186654ea" gracePeriod=30 Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.203676 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g9g6w"] Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.204113 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-g9g6w" podUID="741767ab-ed9f-4b68-83bb-41aa25497741" containerName="registry-server" containerID="cri-o://5e4989257e5025288275554e3c16052be2bfdc1300fb33d2662dbd9046884c97" gracePeriod=30 Dec 09 09:51:10 crc kubenswrapper[4824]: E1209 09:51:10.210173 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5e4989257e5025288275554e3c16052be2bfdc1300fb33d2662dbd9046884c97" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 09:51:10 crc kubenswrapper[4824]: E1209 09:51:10.212116 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5e4989257e5025288275554e3c16052be2bfdc1300fb33d2662dbd9046884c97" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 09:51:10 crc kubenswrapper[4824]: E1209 09:51:10.217518 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5e4989257e5025288275554e3c16052be2bfdc1300fb33d2662dbd9046884c97" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 09:51:10 crc kubenswrapper[4824]: E1209 09:51:10.217609 4824 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-marketplace/certified-operators-g9g6w" podUID="741767ab-ed9f-4b68-83bb-41aa25497741" containerName="registry-server" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.223581 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zmxrv"] Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.223960 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zmxrv" podUID="e799b94c-33db-4de6-afd4-a8ffc4e9b069" containerName="registry-server" containerID="cri-o://7ffb8642d1d49d8e285914bb967a315745b78b0c10c9a28c4ee530ea14536831" gracePeriod=30 Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.236664 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kv8sn"] Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.237049 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-kv8sn" podUID="54818ac0-7b5e-452f-b2f7-6a87a667a76f" containerName="marketplace-operator" containerID="cri-o://05c140427626f12ded7705e709b041e27d2bc8ca00d4082484def3750ef4f06b" gracePeriod=30 Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.245376 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j7pg7"] Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.245765 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-j7pg7" podUID="58792ed4-e360-4827-bf8c-508146205b78" containerName="registry-server" containerID="cri-o://6cffc534d451ede9615b562a1d9b2686276ce832fde7af2884e85d48b0e45c85" gracePeriod=30 Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.261596 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4tt6w"] Dec 09 09:51:10 crc kubenswrapper[4824]: E1209 09:51:10.262096 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad53b3fe-0e9e-4d56-9ca0-212dde3ac669" containerName="extract-content" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.262121 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad53b3fe-0e9e-4d56-9ca0-212dde3ac669" containerName="extract-content" Dec 09 09:51:10 crc kubenswrapper[4824]: E1209 09:51:10.262140 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad53b3fe-0e9e-4d56-9ca0-212dde3ac669" containerName="extract-utilities" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.262149 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad53b3fe-0e9e-4d56-9ca0-212dde3ac669" containerName="extract-utilities" Dec 09 09:51:10 crc kubenswrapper[4824]: E1209 09:51:10.262203 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e43afd8e-2c79-454d-9c95-a0f65f13dc37" containerName="extract-utilities" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.262213 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e43afd8e-2c79-454d-9c95-a0f65f13dc37" containerName="extract-utilities" Dec 09 09:51:10 crc kubenswrapper[4824]: E1209 09:51:10.262230 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6c3ca8-ba36-4cbc-9143-4f83a611144a" containerName="registry-server" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.262262 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6c3ca8-ba36-4cbc-9143-4f83a611144a" containerName="registry-server" Dec 09 09:51:10 crc kubenswrapper[4824]: E1209 09:51:10.262275 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6c3ca8-ba36-4cbc-9143-4f83a611144a" containerName="extract-utilities" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.262284 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6c3ca8-ba36-4cbc-9143-4f83a611144a" containerName="extract-utilities" Dec 09 09:51:10 crc kubenswrapper[4824]: E1209 09:51:10.262298 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e43afd8e-2c79-454d-9c95-a0f65f13dc37" containerName="extract-content" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.262306 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e43afd8e-2c79-454d-9c95-a0f65f13dc37" containerName="extract-content" Dec 09 09:51:10 crc kubenswrapper[4824]: E1209 09:51:10.262335 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad53b3fe-0e9e-4d56-9ca0-212dde3ac669" containerName="registry-server" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.262344 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad53b3fe-0e9e-4d56-9ca0-212dde3ac669" containerName="registry-server" Dec 09 09:51:10 crc kubenswrapper[4824]: E1209 09:51:10.262355 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e43afd8e-2c79-454d-9c95-a0f65f13dc37" containerName="registry-server" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.262365 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e43afd8e-2c79-454d-9c95-a0f65f13dc37" containerName="registry-server" Dec 09 09:51:10 crc kubenswrapper[4824]: E1209 09:51:10.262382 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d7422fc-b0b4-4118-a0de-980eaf4f4fad" containerName="pruner" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.262390 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d7422fc-b0b4-4118-a0de-980eaf4f4fad" containerName="pruner" Dec 09 09:51:10 crc kubenswrapper[4824]: E1209 09:51:10.262421 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6c3ca8-ba36-4cbc-9143-4f83a611144a" containerName="extract-content" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.262431 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6c3ca8-ba36-4cbc-9143-4f83a611144a" containerName="extract-content" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.262653 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d7422fc-b0b4-4118-a0de-980eaf4f4fad" containerName="pruner" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.262679 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e43afd8e-2c79-454d-9c95-a0f65f13dc37" containerName="registry-server" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.262694 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a6c3ca8-ba36-4cbc-9143-4f83a611144a" containerName="registry-server" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.262708 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad53b3fe-0e9e-4d56-9ca0-212dde3ac669" containerName="registry-server" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.263542 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4tt6w" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.268431 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qpbgg"] Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.268902 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qpbgg" podUID="3715eab2-b516-4d39-a23f-664a70efe5b8" containerName="registry-server" containerID="cri-o://2e92fd06942e02ae8b3e4a61127f817724523a7371da5f756c925e65a0db72a8" gracePeriod=30 Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.282398 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4tt6w"] Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.327417 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1b70aae4-4b6b-457c-a358-0f7ce11b8206-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4tt6w\" (UID: \"1b70aae4-4b6b-457c-a358-0f7ce11b8206\") " pod="openshift-marketplace/marketplace-operator-79b997595-4tt6w" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.327515 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxkhl\" (UniqueName: \"kubernetes.io/projected/1b70aae4-4b6b-457c-a358-0f7ce11b8206-kube-api-access-jxkhl\") pod \"marketplace-operator-79b997595-4tt6w\" (UID: \"1b70aae4-4b6b-457c-a358-0f7ce11b8206\") " pod="openshift-marketplace/marketplace-operator-79b997595-4tt6w" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.327562 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b70aae4-4b6b-457c-a358-0f7ce11b8206-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4tt6w\" (UID: \"1b70aae4-4b6b-457c-a358-0f7ce11b8206\") " pod="openshift-marketplace/marketplace-operator-79b997595-4tt6w" Dec 09 09:51:10 crc kubenswrapper[4824]: E1209 09:51:10.386884 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5e4989257e5025288275554e3c16052be2bfdc1300fb33d2662dbd9046884c97 is running failed: container process not found" containerID="5e4989257e5025288275554e3c16052be2bfdc1300fb33d2662dbd9046884c97" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 09:51:10 crc kubenswrapper[4824]: E1209 09:51:10.387757 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5e4989257e5025288275554e3c16052be2bfdc1300fb33d2662dbd9046884c97 is running failed: container process not found" containerID="5e4989257e5025288275554e3c16052be2bfdc1300fb33d2662dbd9046884c97" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 09:51:10 crc kubenswrapper[4824]: E1209 09:51:10.388132 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5e4989257e5025288275554e3c16052be2bfdc1300fb33d2662dbd9046884c97 is running failed: container process not found" containerID="5e4989257e5025288275554e3c16052be2bfdc1300fb33d2662dbd9046884c97" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 09:51:10 crc kubenswrapper[4824]: E1209 09:51:10.388186 4824 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5e4989257e5025288275554e3c16052be2bfdc1300fb33d2662dbd9046884c97 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-g9g6w" podUID="741767ab-ed9f-4b68-83bb-41aa25497741" containerName="registry-server" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.429323 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxkhl\" (UniqueName: \"kubernetes.io/projected/1b70aae4-4b6b-457c-a358-0f7ce11b8206-kube-api-access-jxkhl\") pod \"marketplace-operator-79b997595-4tt6w\" (UID: \"1b70aae4-4b6b-457c-a358-0f7ce11b8206\") " pod="openshift-marketplace/marketplace-operator-79b997595-4tt6w" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.429405 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b70aae4-4b6b-457c-a358-0f7ce11b8206-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4tt6w\" (UID: \"1b70aae4-4b6b-457c-a358-0f7ce11b8206\") " pod="openshift-marketplace/marketplace-operator-79b997595-4tt6w" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.429487 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1b70aae4-4b6b-457c-a358-0f7ce11b8206-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4tt6w\" (UID: \"1b70aae4-4b6b-457c-a358-0f7ce11b8206\") " pod="openshift-marketplace/marketplace-operator-79b997595-4tt6w" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.433647 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b70aae4-4b6b-457c-a358-0f7ce11b8206-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4tt6w\" (UID: \"1b70aae4-4b6b-457c-a358-0f7ce11b8206\") " pod="openshift-marketplace/marketplace-operator-79b997595-4tt6w" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.438906 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1b70aae4-4b6b-457c-a358-0f7ce11b8206-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4tt6w\" (UID: \"1b70aae4-4b6b-457c-a358-0f7ce11b8206\") " pod="openshift-marketplace/marketplace-operator-79b997595-4tt6w" Dec 09 09:51:10 crc kubenswrapper[4824]: E1209 09:51:10.441306 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7ffb8642d1d49d8e285914bb967a315745b78b0c10c9a28c4ee530ea14536831 is running failed: container process not found" containerID="7ffb8642d1d49d8e285914bb967a315745b78b0c10c9a28c4ee530ea14536831" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 09:51:10 crc kubenswrapper[4824]: E1209 09:51:10.446079 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7ffb8642d1d49d8e285914bb967a315745b78b0c10c9a28c4ee530ea14536831 is running failed: container process not found" containerID="7ffb8642d1d49d8e285914bb967a315745b78b0c10c9a28c4ee530ea14536831" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.451959 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxkhl\" (UniqueName: \"kubernetes.io/projected/1b70aae4-4b6b-457c-a358-0f7ce11b8206-kube-api-access-jxkhl\") pod \"marketplace-operator-79b997595-4tt6w\" (UID: \"1b70aae4-4b6b-457c-a358-0f7ce11b8206\") " pod="openshift-marketplace/marketplace-operator-79b997595-4tt6w" Dec 09 09:51:10 crc kubenswrapper[4824]: E1209 09:51:10.455435 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7ffb8642d1d49d8e285914bb967a315745b78b0c10c9a28c4ee530ea14536831 is running failed: container process not found" containerID="7ffb8642d1d49d8e285914bb967a315745b78b0c10c9a28c4ee530ea14536831" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 09:51:10 crc kubenswrapper[4824]: E1209 09:51:10.455561 4824 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7ffb8642d1d49d8e285914bb967a315745b78b0c10c9a28c4ee530ea14536831 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-zmxrv" podUID="e799b94c-33db-4de6-afd4-a8ffc4e9b069" containerName="registry-server" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.585502 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4tt6w" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.609460 4824 generic.go:334] "Generic (PLEG): container finished" podID="e799b94c-33db-4de6-afd4-a8ffc4e9b069" containerID="7ffb8642d1d49d8e285914bb967a315745b78b0c10c9a28c4ee530ea14536831" exitCode=0 Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.609524 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zmxrv" event={"ID":"e799b94c-33db-4de6-afd4-a8ffc4e9b069","Type":"ContainerDied","Data":"7ffb8642d1d49d8e285914bb967a315745b78b0c10c9a28c4ee530ea14536831"} Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.613002 4824 generic.go:334] "Generic (PLEG): container finished" podID="741767ab-ed9f-4b68-83bb-41aa25497741" containerID="5e4989257e5025288275554e3c16052be2bfdc1300fb33d2662dbd9046884c97" exitCode=0 Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.613062 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g9g6w" event={"ID":"741767ab-ed9f-4b68-83bb-41aa25497741","Type":"ContainerDied","Data":"5e4989257e5025288275554e3c16052be2bfdc1300fb33d2662dbd9046884c97"} Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.614319 4824 generic.go:334] "Generic (PLEG): container finished" podID="54818ac0-7b5e-452f-b2f7-6a87a667a76f" containerID="05c140427626f12ded7705e709b041e27d2bc8ca00d4082484def3750ef4f06b" exitCode=0 Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.614383 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kv8sn" event={"ID":"54818ac0-7b5e-452f-b2f7-6a87a667a76f","Type":"ContainerDied","Data":"05c140427626f12ded7705e709b041e27d2bc8ca00d4082484def3750ef4f06b"} Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.622469 4824 generic.go:334] "Generic (PLEG): container finished" podID="db041858-ed93-4df9-a5d2-f79b7f9c3275" containerID="ee4c0468d0910dde9ab8bb7d7aa26531c2e01450ec95e3bfdff6a36b186654ea" exitCode=0 Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.622564 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9pdnx" event={"ID":"db041858-ed93-4df9-a5d2-f79b7f9c3275","Type":"ContainerDied","Data":"ee4c0468d0910dde9ab8bb7d7aa26531c2e01450ec95e3bfdff6a36b186654ea"} Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.627530 4824 generic.go:334] "Generic (PLEG): container finished" podID="58792ed4-e360-4827-bf8c-508146205b78" containerID="6cffc534d451ede9615b562a1d9b2686276ce832fde7af2884e85d48b0e45c85" exitCode=0 Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.627631 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7pg7" event={"ID":"58792ed4-e360-4827-bf8c-508146205b78","Type":"ContainerDied","Data":"6cffc534d451ede9615b562a1d9b2686276ce832fde7af2884e85d48b0e45c85"} Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.632032 4824 generic.go:334] "Generic (PLEG): container finished" podID="3715eab2-b516-4d39-a23f-664a70efe5b8" containerID="2e92fd06942e02ae8b3e4a61127f817724523a7371da5f756c925e65a0db72a8" exitCode=0 Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.632055 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qpbgg" event={"ID":"3715eab2-b516-4d39-a23f-664a70efe5b8","Type":"ContainerDied","Data":"2e92fd06942e02ae8b3e4a61127f817724523a7371da5f756c925e65a0db72a8"} Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.763691 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9pdnx" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.772204 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g9g6w" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.792106 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j7pg7" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.840321 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db041858-ed93-4df9-a5d2-f79b7f9c3275-utilities\") pod \"db041858-ed93-4df9-a5d2-f79b7f9c3275\" (UID: \"db041858-ed93-4df9-a5d2-f79b7f9c3275\") " Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.840907 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wswd5\" (UniqueName: \"kubernetes.io/projected/741767ab-ed9f-4b68-83bb-41aa25497741-kube-api-access-wswd5\") pod \"741767ab-ed9f-4b68-83bb-41aa25497741\" (UID: \"741767ab-ed9f-4b68-83bb-41aa25497741\") " Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.840954 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58792ed4-e360-4827-bf8c-508146205b78-catalog-content\") pod \"58792ed4-e360-4827-bf8c-508146205b78\" (UID: \"58792ed4-e360-4827-bf8c-508146205b78\") " Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.840998 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db041858-ed93-4df9-a5d2-f79b7f9c3275-catalog-content\") pod \"db041858-ed93-4df9-a5d2-f79b7f9c3275\" (UID: \"db041858-ed93-4df9-a5d2-f79b7f9c3275\") " Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.841023 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/741767ab-ed9f-4b68-83bb-41aa25497741-utilities\") pod \"741767ab-ed9f-4b68-83bb-41aa25497741\" (UID: \"741767ab-ed9f-4b68-83bb-41aa25497741\") " Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.841043 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbkw6\" (UniqueName: \"kubernetes.io/projected/db041858-ed93-4df9-a5d2-f79b7f9c3275-kube-api-access-vbkw6\") pod \"db041858-ed93-4df9-a5d2-f79b7f9c3275\" (UID: \"db041858-ed93-4df9-a5d2-f79b7f9c3275\") " Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.841084 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jz54p\" (UniqueName: \"kubernetes.io/projected/58792ed4-e360-4827-bf8c-508146205b78-kube-api-access-jz54p\") pod \"58792ed4-e360-4827-bf8c-508146205b78\" (UID: \"58792ed4-e360-4827-bf8c-508146205b78\") " Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.841115 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58792ed4-e360-4827-bf8c-508146205b78-utilities\") pod \"58792ed4-e360-4827-bf8c-508146205b78\" (UID: \"58792ed4-e360-4827-bf8c-508146205b78\") " Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.841147 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/741767ab-ed9f-4b68-83bb-41aa25497741-catalog-content\") pod \"741767ab-ed9f-4b68-83bb-41aa25497741\" (UID: \"741767ab-ed9f-4b68-83bb-41aa25497741\") " Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.842302 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db041858-ed93-4df9-a5d2-f79b7f9c3275-utilities" (OuterVolumeSpecName: "utilities") pod "db041858-ed93-4df9-a5d2-f79b7f9c3275" (UID: "db041858-ed93-4df9-a5d2-f79b7f9c3275"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.846774 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58792ed4-e360-4827-bf8c-508146205b78-utilities" (OuterVolumeSpecName: "utilities") pod "58792ed4-e360-4827-bf8c-508146205b78" (UID: "58792ed4-e360-4827-bf8c-508146205b78"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.847833 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/741767ab-ed9f-4b68-83bb-41aa25497741-utilities" (OuterVolumeSpecName: "utilities") pod "741767ab-ed9f-4b68-83bb-41aa25497741" (UID: "741767ab-ed9f-4b68-83bb-41aa25497741"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.850152 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/741767ab-ed9f-4b68-83bb-41aa25497741-kube-api-access-wswd5" (OuterVolumeSpecName: "kube-api-access-wswd5") pod "741767ab-ed9f-4b68-83bb-41aa25497741" (UID: "741767ab-ed9f-4b68-83bb-41aa25497741"). InnerVolumeSpecName "kube-api-access-wswd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.857871 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db041858-ed93-4df9-a5d2-f79b7f9c3275-kube-api-access-vbkw6" (OuterVolumeSpecName: "kube-api-access-vbkw6") pod "db041858-ed93-4df9-a5d2-f79b7f9c3275" (UID: "db041858-ed93-4df9-a5d2-f79b7f9c3275"). InnerVolumeSpecName "kube-api-access-vbkw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.861977 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58792ed4-e360-4827-bf8c-508146205b78-kube-api-access-jz54p" (OuterVolumeSpecName: "kube-api-access-jz54p") pod "58792ed4-e360-4827-bf8c-508146205b78" (UID: "58792ed4-e360-4827-bf8c-508146205b78"). InnerVolumeSpecName "kube-api-access-jz54p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.919184 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58792ed4-e360-4827-bf8c-508146205b78-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "58792ed4-e360-4827-bf8c-508146205b78" (UID: "58792ed4-e360-4827-bf8c-508146205b78"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.922577 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kv8sn" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.933238 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/741767ab-ed9f-4b68-83bb-41aa25497741-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "741767ab-ed9f-4b68-83bb-41aa25497741" (UID: "741767ab-ed9f-4b68-83bb-41aa25497741"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.941593 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qpbgg" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.943053 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/741767ab-ed9f-4b68-83bb-41aa25497741-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.943099 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbkw6\" (UniqueName: \"kubernetes.io/projected/db041858-ed93-4df9-a5d2-f79b7f9c3275-kube-api-access-vbkw6\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.943115 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jz54p\" (UniqueName: \"kubernetes.io/projected/58792ed4-e360-4827-bf8c-508146205b78-kube-api-access-jz54p\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.943184 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58792ed4-e360-4827-bf8c-508146205b78-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.943195 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/741767ab-ed9f-4b68-83bb-41aa25497741-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.943204 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db041858-ed93-4df9-a5d2-f79b7f9c3275-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.943217 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wswd5\" (UniqueName: \"kubernetes.io/projected/741767ab-ed9f-4b68-83bb-41aa25497741-kube-api-access-wswd5\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.943227 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58792ed4-e360-4827-bf8c-508146205b78-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.951173 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4tt6w"] Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.956420 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db041858-ed93-4df9-a5d2-f79b7f9c3275-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "db041858-ed93-4df9-a5d2-f79b7f9c3275" (UID: "db041858-ed93-4df9-a5d2-f79b7f9c3275"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:51:10 crc kubenswrapper[4824]: W1209 09:51:10.966447 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b70aae4_4b6b_457c_a358_0f7ce11b8206.slice/crio-9634550803c2feea1c0c6ad87f5ece2a5736af9df9cdc3926d1962580537a675 WatchSource:0}: Error finding container 9634550803c2feea1c0c6ad87f5ece2a5736af9df9cdc3926d1962580537a675: Status 404 returned error can't find the container with id 9634550803c2feea1c0c6ad87f5ece2a5736af9df9cdc3926d1962580537a675 Dec 09 09:51:10 crc kubenswrapper[4824]: I1209 09:51:10.967536 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zmxrv" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.048723 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db041858-ed93-4df9-a5d2-f79b7f9c3275-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.150444 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e799b94c-33db-4de6-afd4-a8ffc4e9b069-utilities\") pod \"e799b94c-33db-4de6-afd4-a8ffc4e9b069\" (UID: \"e799b94c-33db-4de6-afd4-a8ffc4e9b069\") " Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.150905 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3715eab2-b516-4d39-a23f-664a70efe5b8-utilities\") pod \"3715eab2-b516-4d39-a23f-664a70efe5b8\" (UID: \"3715eab2-b516-4d39-a23f-664a70efe5b8\") " Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.150951 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5qsr\" (UniqueName: \"kubernetes.io/projected/3715eab2-b516-4d39-a23f-664a70efe5b8-kube-api-access-b5qsr\") pod \"3715eab2-b516-4d39-a23f-664a70efe5b8\" (UID: \"3715eab2-b516-4d39-a23f-664a70efe5b8\") " Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.151008 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/54818ac0-7b5e-452f-b2f7-6a87a667a76f-marketplace-trusted-ca\") pod \"54818ac0-7b5e-452f-b2f7-6a87a667a76f\" (UID: \"54818ac0-7b5e-452f-b2f7-6a87a667a76f\") " Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.151049 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e799b94c-33db-4de6-afd4-a8ffc4e9b069-catalog-content\") pod \"e799b94c-33db-4de6-afd4-a8ffc4e9b069\" (UID: \"e799b94c-33db-4de6-afd4-a8ffc4e9b069\") " Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.151084 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28b6h\" (UniqueName: \"kubernetes.io/projected/e799b94c-33db-4de6-afd4-a8ffc4e9b069-kube-api-access-28b6h\") pod \"e799b94c-33db-4de6-afd4-a8ffc4e9b069\" (UID: \"e799b94c-33db-4de6-afd4-a8ffc4e9b069\") " Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.151122 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3715eab2-b516-4d39-a23f-664a70efe5b8-catalog-content\") pod \"3715eab2-b516-4d39-a23f-664a70efe5b8\" (UID: \"3715eab2-b516-4d39-a23f-664a70efe5b8\") " Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.151243 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2wk8\" (UniqueName: \"kubernetes.io/projected/54818ac0-7b5e-452f-b2f7-6a87a667a76f-kube-api-access-n2wk8\") pod \"54818ac0-7b5e-452f-b2f7-6a87a667a76f\" (UID: \"54818ac0-7b5e-452f-b2f7-6a87a667a76f\") " Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.151302 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/54818ac0-7b5e-452f-b2f7-6a87a667a76f-marketplace-operator-metrics\") pod \"54818ac0-7b5e-452f-b2f7-6a87a667a76f\" (UID: \"54818ac0-7b5e-452f-b2f7-6a87a667a76f\") " Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.151456 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e799b94c-33db-4de6-afd4-a8ffc4e9b069-utilities" (OuterVolumeSpecName: "utilities") pod "e799b94c-33db-4de6-afd4-a8ffc4e9b069" (UID: "e799b94c-33db-4de6-afd4-a8ffc4e9b069"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.152070 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e799b94c-33db-4de6-afd4-a8ffc4e9b069-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.152566 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54818ac0-7b5e-452f-b2f7-6a87a667a76f-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "54818ac0-7b5e-452f-b2f7-6a87a667a76f" (UID: "54818ac0-7b5e-452f-b2f7-6a87a667a76f"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.153477 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3715eab2-b516-4d39-a23f-664a70efe5b8-utilities" (OuterVolumeSpecName: "utilities") pod "3715eab2-b516-4d39-a23f-664a70efe5b8" (UID: "3715eab2-b516-4d39-a23f-664a70efe5b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.156284 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e799b94c-33db-4de6-afd4-a8ffc4e9b069-kube-api-access-28b6h" (OuterVolumeSpecName: "kube-api-access-28b6h") pod "e799b94c-33db-4de6-afd4-a8ffc4e9b069" (UID: "e799b94c-33db-4de6-afd4-a8ffc4e9b069"). InnerVolumeSpecName "kube-api-access-28b6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.156771 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3715eab2-b516-4d39-a23f-664a70efe5b8-kube-api-access-b5qsr" (OuterVolumeSpecName: "kube-api-access-b5qsr") pod "3715eab2-b516-4d39-a23f-664a70efe5b8" (UID: "3715eab2-b516-4d39-a23f-664a70efe5b8"). InnerVolumeSpecName "kube-api-access-b5qsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.157236 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54818ac0-7b5e-452f-b2f7-6a87a667a76f-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "54818ac0-7b5e-452f-b2f7-6a87a667a76f" (UID: "54818ac0-7b5e-452f-b2f7-6a87a667a76f"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.157496 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54818ac0-7b5e-452f-b2f7-6a87a667a76f-kube-api-access-n2wk8" (OuterVolumeSpecName: "kube-api-access-n2wk8") pod "54818ac0-7b5e-452f-b2f7-6a87a667a76f" (UID: "54818ac0-7b5e-452f-b2f7-6a87a667a76f"). InnerVolumeSpecName "kube-api-access-n2wk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.233572 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e799b94c-33db-4de6-afd4-a8ffc4e9b069-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e799b94c-33db-4de6-afd4-a8ffc4e9b069" (UID: "e799b94c-33db-4de6-afd4-a8ffc4e9b069"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.252716 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2wk8\" (UniqueName: \"kubernetes.io/projected/54818ac0-7b5e-452f-b2f7-6a87a667a76f-kube-api-access-n2wk8\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.252764 4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/54818ac0-7b5e-452f-b2f7-6a87a667a76f-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.252793 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3715eab2-b516-4d39-a23f-664a70efe5b8-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.252811 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5qsr\" (UniqueName: \"kubernetes.io/projected/3715eab2-b516-4d39-a23f-664a70efe5b8-kube-api-access-b5qsr\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.252827 4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/54818ac0-7b5e-452f-b2f7-6a87a667a76f-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.252844 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e799b94c-33db-4de6-afd4-a8ffc4e9b069-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.252857 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28b6h\" (UniqueName: \"kubernetes.io/projected/e799b94c-33db-4de6-afd4-a8ffc4e9b069-kube-api-access-28b6h\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.279009 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3715eab2-b516-4d39-a23f-664a70efe5b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3715eab2-b516-4d39-a23f-664a70efe5b8" (UID: "3715eab2-b516-4d39-a23f-664a70efe5b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.353906 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3715eab2-b516-4d39-a23f-664a70efe5b8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.641024 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g9g6w" event={"ID":"741767ab-ed9f-4b68-83bb-41aa25497741","Type":"ContainerDied","Data":"f33e5882217416bbc7a93a824c8b1bb99d2c6a6b81b5027edaa71d9882bccf22"} Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.641107 4824 scope.go:117] "RemoveContainer" containerID="5e4989257e5025288275554e3c16052be2bfdc1300fb33d2662dbd9046884c97" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.641058 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g9g6w" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.645379 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4tt6w" event={"ID":"1b70aae4-4b6b-457c-a358-0f7ce11b8206","Type":"ContainerStarted","Data":"effe8a4c4c55f5a20246c1c3eee827025019fc93cdca0ee47653703b1758cc91"} Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.645467 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4tt6w" event={"ID":"1b70aae4-4b6b-457c-a358-0f7ce11b8206","Type":"ContainerStarted","Data":"9634550803c2feea1c0c6ad87f5ece2a5736af9df9cdc3926d1962580537a675"} Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.647856 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4tt6w" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.649925 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kv8sn" event={"ID":"54818ac0-7b5e-452f-b2f7-6a87a667a76f","Type":"ContainerDied","Data":"9c2c0529106a07b565471232591095d533d8728797d7f5fe4875267115162a8b"} Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.650017 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kv8sn" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.652646 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9pdnx" event={"ID":"db041858-ed93-4df9-a5d2-f79b7f9c3275","Type":"ContainerDied","Data":"b8ce610308575d8c5358c5bea7b4c7667f271d057920f405b1d3c55dcf210509"} Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.652892 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9pdnx" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.655336 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-4tt6w" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.660677 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7pg7" event={"ID":"58792ed4-e360-4827-bf8c-508146205b78","Type":"ContainerDied","Data":"a9e9ab1dd2ff7148955bea103166ef1a37504cb02d57cf81b810e0803afacd7e"} Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.660742 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j7pg7" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.664395 4824 scope.go:117] "RemoveContainer" containerID="082ab047d0517819a6cf6b09c30087eca66647f8a9d48e6777e39d727ced3678" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.665054 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qpbgg" event={"ID":"3715eab2-b516-4d39-a23f-664a70efe5b8","Type":"ContainerDied","Data":"96955b17111cd0bbe43ca287387c357b0877f0951747ea335b834b8749f93d43"} Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.665856 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qpbgg" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.673072 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zmxrv" event={"ID":"e799b94c-33db-4de6-afd4-a8ffc4e9b069","Type":"ContainerDied","Data":"0b490d86ca847167731a7a3b93a1f9cdbdc5cda854d96971826a82bd616519b0"} Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.673155 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zmxrv" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.684568 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-4tt6w" podStartSLOduration=1.684536772 podStartE2EDuration="1.684536772s" podCreationTimestamp="2025-12-09 09:51:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:51:11.677843304 +0000 UTC m=+228.012347971" watchObservedRunningTime="2025-12-09 09:51:11.684536772 +0000 UTC m=+228.019041439" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.687944 4824 scope.go:117] "RemoveContainer" containerID="08e14507f9df56ef50a945ee72ab5c6c72f1ea4b9d3e82f51755ec5f98db9954" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.723430 4824 scope.go:117] "RemoveContainer" containerID="05c140427626f12ded7705e709b041e27d2bc8ca00d4082484def3750ef4f06b" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.753628 4824 scope.go:117] "RemoveContainer" containerID="ee4c0468d0910dde9ab8bb7d7aa26531c2e01450ec95e3bfdff6a36b186654ea" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.768736 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g9g6w"] Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.777490 4824 scope.go:117] "RemoveContainer" containerID="958189bac093637292ef3b226b1e67ef6dc51a7d5dd646cbb02881822d51ecfb" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.782141 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-g9g6w"] Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.798743 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9pdnx"] Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.800415 4824 scope.go:117] "RemoveContainer" containerID="2962bf05684f13cc5e6a973a03e16b20b1b2afcf9b7dba35c04ac979fdf1a4bb" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.812080 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9pdnx"] Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.817962 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j7pg7"] Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.826981 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-j7pg7"] Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.833918 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zmxrv"] Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.838276 4824 scope.go:117] "RemoveContainer" containerID="6cffc534d451ede9615b562a1d9b2686276ce832fde7af2884e85d48b0e45c85" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.838508 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zmxrv"] Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.842944 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kv8sn"] Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.852817 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kv8sn"] Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.856849 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qpbgg"] Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.858732 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qpbgg"] Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.863173 4824 scope.go:117] "RemoveContainer" containerID="3bd296d6db16386b94a4d82a7b737b3623c71538b2dbc27f60b5e97b616fca4a" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.880619 4824 scope.go:117] "RemoveContainer" containerID="df6bde8f2b4593a55bca46d073d569cdf13cae123f4bc5d63c2e2932a77affe5" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.898866 4824 scope.go:117] "RemoveContainer" containerID="2e92fd06942e02ae8b3e4a61127f817724523a7371da5f756c925e65a0db72a8" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.916912 4824 scope.go:117] "RemoveContainer" containerID="1393f881c6354f5700898a5abe61086d31f974db1a007d8e11b201dad1e5d3bd" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.920219 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3715eab2-b516-4d39-a23f-664a70efe5b8" path="/var/lib/kubelet/pods/3715eab2-b516-4d39-a23f-664a70efe5b8/volumes" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.920972 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54818ac0-7b5e-452f-b2f7-6a87a667a76f" path="/var/lib/kubelet/pods/54818ac0-7b5e-452f-b2f7-6a87a667a76f/volumes" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.921526 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58792ed4-e360-4827-bf8c-508146205b78" path="/var/lib/kubelet/pods/58792ed4-e360-4827-bf8c-508146205b78/volumes" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.922675 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="741767ab-ed9f-4b68-83bb-41aa25497741" path="/var/lib/kubelet/pods/741767ab-ed9f-4b68-83bb-41aa25497741/volumes" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.923487 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db041858-ed93-4df9-a5d2-f79b7f9c3275" path="/var/lib/kubelet/pods/db041858-ed93-4df9-a5d2-f79b7f9c3275/volumes" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.924697 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e799b94c-33db-4de6-afd4-a8ffc4e9b069" path="/var/lib/kubelet/pods/e799b94c-33db-4de6-afd4-a8ffc4e9b069/volumes" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.940148 4824 scope.go:117] "RemoveContainer" containerID="a938c67634e4808db60c1728f3a58403f72fdb7f5f003adf3973cb3ed2dc52c2" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.953304 4824 scope.go:117] "RemoveContainer" containerID="7ffb8642d1d49d8e285914bb967a315745b78b0c10c9a28c4ee530ea14536831" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.966328 4824 scope.go:117] "RemoveContainer" containerID="f8579e4d3233a7272513eb992802b40a7d632811c573387c5167ef301b317979" Dec 09 09:51:11 crc kubenswrapper[4824]: I1209 09:51:11.985492 4824 scope.go:117] "RemoveContainer" containerID="f4043c535e6fcd55a4841f1ff3aa3e27e70824ff475563a214761e258a197dd6" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.575940 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-582xj"] Dec 09 09:51:12 crc kubenswrapper[4824]: E1209 09:51:12.576689 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db041858-ed93-4df9-a5d2-f79b7f9c3275" containerName="registry-server" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.576708 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="db041858-ed93-4df9-a5d2-f79b7f9c3275" containerName="registry-server" Dec 09 09:51:12 crc kubenswrapper[4824]: E1209 09:51:12.576725 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="741767ab-ed9f-4b68-83bb-41aa25497741" containerName="extract-utilities" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.576734 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="741767ab-ed9f-4b68-83bb-41aa25497741" containerName="extract-utilities" Dec 09 09:51:12 crc kubenswrapper[4824]: E1209 09:51:12.576745 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e799b94c-33db-4de6-afd4-a8ffc4e9b069" containerName="registry-server" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.579056 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e799b94c-33db-4de6-afd4-a8ffc4e9b069" containerName="registry-server" Dec 09 09:51:12 crc kubenswrapper[4824]: E1209 09:51:12.579083 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58792ed4-e360-4827-bf8c-508146205b78" containerName="registry-server" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.579091 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="58792ed4-e360-4827-bf8c-508146205b78" containerName="registry-server" Dec 09 09:51:12 crc kubenswrapper[4824]: E1209 09:51:12.579101 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3715eab2-b516-4d39-a23f-664a70efe5b8" containerName="registry-server" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.579108 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3715eab2-b516-4d39-a23f-664a70efe5b8" containerName="registry-server" Dec 09 09:51:12 crc kubenswrapper[4824]: E1209 09:51:12.579126 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="741767ab-ed9f-4b68-83bb-41aa25497741" containerName="registry-server" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.579133 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="741767ab-ed9f-4b68-83bb-41aa25497741" containerName="registry-server" Dec 09 09:51:12 crc kubenswrapper[4824]: E1209 09:51:12.579140 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58792ed4-e360-4827-bf8c-508146205b78" containerName="extract-content" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.579146 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="58792ed4-e360-4827-bf8c-508146205b78" containerName="extract-content" Dec 09 09:51:12 crc kubenswrapper[4824]: E1209 09:51:12.579155 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e799b94c-33db-4de6-afd4-a8ffc4e9b069" containerName="extract-utilities" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.579161 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e799b94c-33db-4de6-afd4-a8ffc4e9b069" containerName="extract-utilities" Dec 09 09:51:12 crc kubenswrapper[4824]: E1209 09:51:12.579172 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="741767ab-ed9f-4b68-83bb-41aa25497741" containerName="extract-content" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.579178 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="741767ab-ed9f-4b68-83bb-41aa25497741" containerName="extract-content" Dec 09 09:51:12 crc kubenswrapper[4824]: E1209 09:51:12.579190 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3715eab2-b516-4d39-a23f-664a70efe5b8" containerName="extract-content" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.579197 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3715eab2-b516-4d39-a23f-664a70efe5b8" containerName="extract-content" Dec 09 09:51:12 crc kubenswrapper[4824]: E1209 09:51:12.579206 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58792ed4-e360-4827-bf8c-508146205b78" containerName="extract-utilities" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.579212 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="58792ed4-e360-4827-bf8c-508146205b78" containerName="extract-utilities" Dec 09 09:51:12 crc kubenswrapper[4824]: E1209 09:51:12.579223 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db041858-ed93-4df9-a5d2-f79b7f9c3275" containerName="extract-utilities" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.579229 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="db041858-ed93-4df9-a5d2-f79b7f9c3275" containerName="extract-utilities" Dec 09 09:51:12 crc kubenswrapper[4824]: E1209 09:51:12.579237 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3715eab2-b516-4d39-a23f-664a70efe5b8" containerName="extract-utilities" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.579244 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3715eab2-b516-4d39-a23f-664a70efe5b8" containerName="extract-utilities" Dec 09 09:51:12 crc kubenswrapper[4824]: E1209 09:51:12.579250 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db041858-ed93-4df9-a5d2-f79b7f9c3275" containerName="extract-content" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.579256 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="db041858-ed93-4df9-a5d2-f79b7f9c3275" containerName="extract-content" Dec 09 09:51:12 crc kubenswrapper[4824]: E1209 09:51:12.579264 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e799b94c-33db-4de6-afd4-a8ffc4e9b069" containerName="extract-content" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.579270 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e799b94c-33db-4de6-afd4-a8ffc4e9b069" containerName="extract-content" Dec 09 09:51:12 crc kubenswrapper[4824]: E1209 09:51:12.579279 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54818ac0-7b5e-452f-b2f7-6a87a667a76f" containerName="marketplace-operator" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.579285 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="54818ac0-7b5e-452f-b2f7-6a87a667a76f" containerName="marketplace-operator" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.579423 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3715eab2-b516-4d39-a23f-664a70efe5b8" containerName="registry-server" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.579444 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="db041858-ed93-4df9-a5d2-f79b7f9c3275" containerName="registry-server" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.579452 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e799b94c-33db-4de6-afd4-a8ffc4e9b069" containerName="registry-server" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.579459 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="54818ac0-7b5e-452f-b2f7-6a87a667a76f" containerName="marketplace-operator" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.579466 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="741767ab-ed9f-4b68-83bb-41aa25497741" containerName="registry-server" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.579475 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="58792ed4-e360-4827-bf8c-508146205b78" containerName="registry-server" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.580516 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-582xj" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.584382 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.588939 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-582xj"] Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.776271 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2572cf3-89f1-49d7-9246-9350703e9764-utilities\") pod \"redhat-operators-582xj\" (UID: \"f2572cf3-89f1-49d7-9246-9350703e9764\") " pod="openshift-marketplace/redhat-operators-582xj" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.776377 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfp67\" (UniqueName: \"kubernetes.io/projected/f2572cf3-89f1-49d7-9246-9350703e9764-kube-api-access-hfp67\") pod \"redhat-operators-582xj\" (UID: \"f2572cf3-89f1-49d7-9246-9350703e9764\") " pod="openshift-marketplace/redhat-operators-582xj" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.776434 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2572cf3-89f1-49d7-9246-9350703e9764-catalog-content\") pod \"redhat-operators-582xj\" (UID: \"f2572cf3-89f1-49d7-9246-9350703e9764\") " pod="openshift-marketplace/redhat-operators-582xj" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.878085 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2572cf3-89f1-49d7-9246-9350703e9764-utilities\") pod \"redhat-operators-582xj\" (UID: \"f2572cf3-89f1-49d7-9246-9350703e9764\") " pod="openshift-marketplace/redhat-operators-582xj" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.878225 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfp67\" (UniqueName: \"kubernetes.io/projected/f2572cf3-89f1-49d7-9246-9350703e9764-kube-api-access-hfp67\") pod \"redhat-operators-582xj\" (UID: \"f2572cf3-89f1-49d7-9246-9350703e9764\") " pod="openshift-marketplace/redhat-operators-582xj" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.878273 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2572cf3-89f1-49d7-9246-9350703e9764-catalog-content\") pod \"redhat-operators-582xj\" (UID: \"f2572cf3-89f1-49d7-9246-9350703e9764\") " pod="openshift-marketplace/redhat-operators-582xj" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.879020 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2572cf3-89f1-49d7-9246-9350703e9764-utilities\") pod \"redhat-operators-582xj\" (UID: \"f2572cf3-89f1-49d7-9246-9350703e9764\") " pod="openshift-marketplace/redhat-operators-582xj" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.879064 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2572cf3-89f1-49d7-9246-9350703e9764-catalog-content\") pod \"redhat-operators-582xj\" (UID: \"f2572cf3-89f1-49d7-9246-9350703e9764\") " pod="openshift-marketplace/redhat-operators-582xj" Dec 09 09:51:12 crc kubenswrapper[4824]: I1209 09:51:12.909092 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfp67\" (UniqueName: \"kubernetes.io/projected/f2572cf3-89f1-49d7-9246-9350703e9764-kube-api-access-hfp67\") pod \"redhat-operators-582xj\" (UID: \"f2572cf3-89f1-49d7-9246-9350703e9764\") " pod="openshift-marketplace/redhat-operators-582xj" Dec 09 09:51:13 crc kubenswrapper[4824]: I1209 09:51:13.167285 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-95jfr"] Dec 09 09:51:13 crc kubenswrapper[4824]: I1209 09:51:13.169043 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-95jfr" Dec 09 09:51:13 crc kubenswrapper[4824]: I1209 09:51:13.172151 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 09 09:51:13 crc kubenswrapper[4824]: I1209 09:51:13.180238 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-95jfr"] Dec 09 09:51:13 crc kubenswrapper[4824]: I1209 09:51:13.182717 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flk4t\" (UniqueName: \"kubernetes.io/projected/6ffd99b7-9bb8-479e-a1a1-7f31ee006e38-kube-api-access-flk4t\") pod \"redhat-marketplace-95jfr\" (UID: \"6ffd99b7-9bb8-479e-a1a1-7f31ee006e38\") " pod="openshift-marketplace/redhat-marketplace-95jfr" Dec 09 09:51:13 crc kubenswrapper[4824]: I1209 09:51:13.182819 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ffd99b7-9bb8-479e-a1a1-7f31ee006e38-utilities\") pod \"redhat-marketplace-95jfr\" (UID: \"6ffd99b7-9bb8-479e-a1a1-7f31ee006e38\") " pod="openshift-marketplace/redhat-marketplace-95jfr" Dec 09 09:51:13 crc kubenswrapper[4824]: I1209 09:51:13.182886 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ffd99b7-9bb8-479e-a1a1-7f31ee006e38-catalog-content\") pod \"redhat-marketplace-95jfr\" (UID: \"6ffd99b7-9bb8-479e-a1a1-7f31ee006e38\") " pod="openshift-marketplace/redhat-marketplace-95jfr" Dec 09 09:51:13 crc kubenswrapper[4824]: I1209 09:51:13.201556 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-582xj" Dec 09 09:51:13 crc kubenswrapper[4824]: I1209 09:51:13.284654 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ffd99b7-9bb8-479e-a1a1-7f31ee006e38-utilities\") pod \"redhat-marketplace-95jfr\" (UID: \"6ffd99b7-9bb8-479e-a1a1-7f31ee006e38\") " pod="openshift-marketplace/redhat-marketplace-95jfr" Dec 09 09:51:13 crc kubenswrapper[4824]: I1209 09:51:13.284762 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ffd99b7-9bb8-479e-a1a1-7f31ee006e38-catalog-content\") pod \"redhat-marketplace-95jfr\" (UID: \"6ffd99b7-9bb8-479e-a1a1-7f31ee006e38\") " pod="openshift-marketplace/redhat-marketplace-95jfr" Dec 09 09:51:13 crc kubenswrapper[4824]: I1209 09:51:13.284864 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flk4t\" (UniqueName: \"kubernetes.io/projected/6ffd99b7-9bb8-479e-a1a1-7f31ee006e38-kube-api-access-flk4t\") pod \"redhat-marketplace-95jfr\" (UID: \"6ffd99b7-9bb8-479e-a1a1-7f31ee006e38\") " pod="openshift-marketplace/redhat-marketplace-95jfr" Dec 09 09:51:13 crc kubenswrapper[4824]: I1209 09:51:13.285402 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ffd99b7-9bb8-479e-a1a1-7f31ee006e38-catalog-content\") pod \"redhat-marketplace-95jfr\" (UID: \"6ffd99b7-9bb8-479e-a1a1-7f31ee006e38\") " pod="openshift-marketplace/redhat-marketplace-95jfr" Dec 09 09:51:13 crc kubenswrapper[4824]: I1209 09:51:13.285447 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ffd99b7-9bb8-479e-a1a1-7f31ee006e38-utilities\") pod \"redhat-marketplace-95jfr\" (UID: \"6ffd99b7-9bb8-479e-a1a1-7f31ee006e38\") " pod="openshift-marketplace/redhat-marketplace-95jfr" Dec 09 09:51:13 crc kubenswrapper[4824]: I1209 09:51:13.307099 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flk4t\" (UniqueName: \"kubernetes.io/projected/6ffd99b7-9bb8-479e-a1a1-7f31ee006e38-kube-api-access-flk4t\") pod \"redhat-marketplace-95jfr\" (UID: \"6ffd99b7-9bb8-479e-a1a1-7f31ee006e38\") " pod="openshift-marketplace/redhat-marketplace-95jfr" Dec 09 09:51:13 crc kubenswrapper[4824]: I1209 09:51:13.565400 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-95jfr" Dec 09 09:51:13 crc kubenswrapper[4824]: I1209 09:51:13.650477 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-582xj"] Dec 09 09:51:13 crc kubenswrapper[4824]: I1209 09:51:13.719406 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-582xj" event={"ID":"f2572cf3-89f1-49d7-9246-9350703e9764","Type":"ContainerStarted","Data":"2b6d728b1616c60d2818765f77a174c0d9080b18f44fe9847df80315a1ce03b7"} Dec 09 09:51:14 crc kubenswrapper[4824]: W1209 09:51:14.014928 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ffd99b7_9bb8_479e_a1a1_7f31ee006e38.slice/crio-40426736e5ec987c202fdb18d252ac5ab997a2fe7fd043bda65272ec850e6ec8 WatchSource:0}: Error finding container 40426736e5ec987c202fdb18d252ac5ab997a2fe7fd043bda65272ec850e6ec8: Status 404 returned error can't find the container with id 40426736e5ec987c202fdb18d252ac5ab997a2fe7fd043bda65272ec850e6ec8 Dec 09 09:51:14 crc kubenswrapper[4824]: I1209 09:51:14.016881 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-95jfr"] Dec 09 09:51:14 crc kubenswrapper[4824]: I1209 09:51:14.728717 4824 generic.go:334] "Generic (PLEG): container finished" podID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" containerID="efe1668989971f1296135ad779203250d322bc536df285decf938a37659c5e41" exitCode=0 Dec 09 09:51:14 crc kubenswrapper[4824]: I1209 09:51:14.728809 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-95jfr" event={"ID":"6ffd99b7-9bb8-479e-a1a1-7f31ee006e38","Type":"ContainerDied","Data":"efe1668989971f1296135ad779203250d322bc536df285decf938a37659c5e41"} Dec 09 09:51:14 crc kubenswrapper[4824]: I1209 09:51:14.728874 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-95jfr" event={"ID":"6ffd99b7-9bb8-479e-a1a1-7f31ee006e38","Type":"ContainerStarted","Data":"40426736e5ec987c202fdb18d252ac5ab997a2fe7fd043bda65272ec850e6ec8"} Dec 09 09:51:14 crc kubenswrapper[4824]: I1209 09:51:14.733288 4824 generic.go:334] "Generic (PLEG): container finished" podID="f2572cf3-89f1-49d7-9246-9350703e9764" containerID="13cf4bb8b57e2adb0373aefde94c2f50c8e51a8eadf441262210f48103d46704" exitCode=0 Dec 09 09:51:14 crc kubenswrapper[4824]: I1209 09:51:14.733365 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-582xj" event={"ID":"f2572cf3-89f1-49d7-9246-9350703e9764","Type":"ContainerDied","Data":"13cf4bb8b57e2adb0373aefde94c2f50c8e51a8eadf441262210f48103d46704"} Dec 09 09:51:14 crc kubenswrapper[4824]: I1209 09:51:14.971533 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jm5ld"] Dec 09 09:51:14 crc kubenswrapper[4824]: I1209 09:51:14.972588 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jm5ld" Dec 09 09:51:14 crc kubenswrapper[4824]: I1209 09:51:14.977663 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 09 09:51:14 crc kubenswrapper[4824]: I1209 09:51:14.990255 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jm5ld"] Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.112932 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc5pn\" (UniqueName: \"kubernetes.io/projected/bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271-kube-api-access-zc5pn\") pod \"community-operators-jm5ld\" (UID: \"bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271\") " pod="openshift-marketplace/community-operators-jm5ld" Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.113125 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271-utilities\") pod \"community-operators-jm5ld\" (UID: \"bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271\") " pod="openshift-marketplace/community-operators-jm5ld" Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.113151 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271-catalog-content\") pod \"community-operators-jm5ld\" (UID: \"bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271\") " pod="openshift-marketplace/community-operators-jm5ld" Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.214472 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271-utilities\") pod \"community-operators-jm5ld\" (UID: \"bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271\") " pod="openshift-marketplace/community-operators-jm5ld" Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.214567 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271-catalog-content\") pod \"community-operators-jm5ld\" (UID: \"bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271\") " pod="openshift-marketplace/community-operators-jm5ld" Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.214645 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc5pn\" (UniqueName: \"kubernetes.io/projected/bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271-kube-api-access-zc5pn\") pod \"community-operators-jm5ld\" (UID: \"bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271\") " pod="openshift-marketplace/community-operators-jm5ld" Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.215125 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271-utilities\") pod \"community-operators-jm5ld\" (UID: \"bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271\") " pod="openshift-marketplace/community-operators-jm5ld" Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.215147 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271-catalog-content\") pod \"community-operators-jm5ld\" (UID: \"bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271\") " pod="openshift-marketplace/community-operators-jm5ld" Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.237808 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc5pn\" (UniqueName: \"kubernetes.io/projected/bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271-kube-api-access-zc5pn\") pod \"community-operators-jm5ld\" (UID: \"bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271\") " pod="openshift-marketplace/community-operators-jm5ld" Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.303966 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jm5ld" Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.566717 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2cbmd"] Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.570400 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2cbmd" Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.576009 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.576880 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2cbmd"] Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.722480 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rljmv\" (UniqueName: \"kubernetes.io/projected/257d6d77-75f1-41e0-8bd2-c6ec3c1ff582-kube-api-access-rljmv\") pod \"certified-operators-2cbmd\" (UID: \"257d6d77-75f1-41e0-8bd2-c6ec3c1ff582\") " pod="openshift-marketplace/certified-operators-2cbmd" Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.722583 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/257d6d77-75f1-41e0-8bd2-c6ec3c1ff582-catalog-content\") pod \"certified-operators-2cbmd\" (UID: \"257d6d77-75f1-41e0-8bd2-c6ec3c1ff582\") " pod="openshift-marketplace/certified-operators-2cbmd" Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.722642 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/257d6d77-75f1-41e0-8bd2-c6ec3c1ff582-utilities\") pod \"certified-operators-2cbmd\" (UID: \"257d6d77-75f1-41e0-8bd2-c6ec3c1ff582\") " pod="openshift-marketplace/certified-operators-2cbmd" Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.765245 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jm5ld"] Dec 09 09:51:15 crc kubenswrapper[4824]: W1209 09:51:15.772425 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb0311bb_7c1c_43a6_aac3_8d6dcdcf6271.slice/crio-a4632c8846e6d2596d390713d5f8907592d634e3f47a6c2c002e6b3fe035806c WatchSource:0}: Error finding container a4632c8846e6d2596d390713d5f8907592d634e3f47a6c2c002e6b3fe035806c: Status 404 returned error can't find the container with id a4632c8846e6d2596d390713d5f8907592d634e3f47a6c2c002e6b3fe035806c Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.823560 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/257d6d77-75f1-41e0-8bd2-c6ec3c1ff582-catalog-content\") pod \"certified-operators-2cbmd\" (UID: \"257d6d77-75f1-41e0-8bd2-c6ec3c1ff582\") " pod="openshift-marketplace/certified-operators-2cbmd" Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.823632 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/257d6d77-75f1-41e0-8bd2-c6ec3c1ff582-utilities\") pod \"certified-operators-2cbmd\" (UID: \"257d6d77-75f1-41e0-8bd2-c6ec3c1ff582\") " pod="openshift-marketplace/certified-operators-2cbmd" Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.823683 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rljmv\" (UniqueName: \"kubernetes.io/projected/257d6d77-75f1-41e0-8bd2-c6ec3c1ff582-kube-api-access-rljmv\") pod \"certified-operators-2cbmd\" (UID: \"257d6d77-75f1-41e0-8bd2-c6ec3c1ff582\") " pod="openshift-marketplace/certified-operators-2cbmd" Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.824335 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/257d6d77-75f1-41e0-8bd2-c6ec3c1ff582-catalog-content\") pod \"certified-operators-2cbmd\" (UID: \"257d6d77-75f1-41e0-8bd2-c6ec3c1ff582\") " pod="openshift-marketplace/certified-operators-2cbmd" Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.824513 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/257d6d77-75f1-41e0-8bd2-c6ec3c1ff582-utilities\") pod \"certified-operators-2cbmd\" (UID: \"257d6d77-75f1-41e0-8bd2-c6ec3c1ff582\") " pod="openshift-marketplace/certified-operators-2cbmd" Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.845258 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rljmv\" (UniqueName: \"kubernetes.io/projected/257d6d77-75f1-41e0-8bd2-c6ec3c1ff582-kube-api-access-rljmv\") pod \"certified-operators-2cbmd\" (UID: \"257d6d77-75f1-41e0-8bd2-c6ec3c1ff582\") " pod="openshift-marketplace/certified-operators-2cbmd" Dec 09 09:51:15 crc kubenswrapper[4824]: I1209 09:51:15.891908 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2cbmd" Dec 09 09:51:16 crc kubenswrapper[4824]: I1209 09:51:16.130774 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2cbmd"] Dec 09 09:51:16 crc kubenswrapper[4824]: W1209 09:51:16.137742 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod257d6d77_75f1_41e0_8bd2_c6ec3c1ff582.slice/crio-644489414a186bb6c76ffb15062547db3c0fbfed1ad7947bfe0b906d5738ec81 WatchSource:0}: Error finding container 644489414a186bb6c76ffb15062547db3c0fbfed1ad7947bfe0b906d5738ec81: Status 404 returned error can't find the container with id 644489414a186bb6c76ffb15062547db3c0fbfed1ad7947bfe0b906d5738ec81 Dec 09 09:51:16 crc kubenswrapper[4824]: I1209 09:51:16.754107 4824 generic.go:334] "Generic (PLEG): container finished" podID="f2572cf3-89f1-49d7-9246-9350703e9764" containerID="b7aaeaea1d5c96a384b4446010d13c651f7ba1788bf6c0499fee8e6d058f5e09" exitCode=0 Dec 09 09:51:16 crc kubenswrapper[4824]: I1209 09:51:16.754188 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-582xj" event={"ID":"f2572cf3-89f1-49d7-9246-9350703e9764","Type":"ContainerDied","Data":"b7aaeaea1d5c96a384b4446010d13c651f7ba1788bf6c0499fee8e6d058f5e09"} Dec 09 09:51:16 crc kubenswrapper[4824]: I1209 09:51:16.759631 4824 generic.go:334] "Generic (PLEG): container finished" podID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" containerID="45005d5f9926188a9f1ff4662f1b49b2f23793ca498faf89a2eea56ba63d5aed" exitCode=0 Dec 09 09:51:16 crc kubenswrapper[4824]: I1209 09:51:16.760145 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cbmd" event={"ID":"257d6d77-75f1-41e0-8bd2-c6ec3c1ff582","Type":"ContainerDied","Data":"45005d5f9926188a9f1ff4662f1b49b2f23793ca498faf89a2eea56ba63d5aed"} Dec 09 09:51:16 crc kubenswrapper[4824]: I1209 09:51:16.760175 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cbmd" event={"ID":"257d6d77-75f1-41e0-8bd2-c6ec3c1ff582","Type":"ContainerStarted","Data":"644489414a186bb6c76ffb15062547db3c0fbfed1ad7947bfe0b906d5738ec81"} Dec 09 09:51:16 crc kubenswrapper[4824]: I1209 09:51:16.766708 4824 generic.go:334] "Generic (PLEG): container finished" podID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" containerID="86447b85ed6f57820778777e25f24724af169a65ee02b662e4a8709e9856f9c0" exitCode=0 Dec 09 09:51:16 crc kubenswrapper[4824]: I1209 09:51:16.766798 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-95jfr" event={"ID":"6ffd99b7-9bb8-479e-a1a1-7f31ee006e38","Type":"ContainerDied","Data":"86447b85ed6f57820778777e25f24724af169a65ee02b662e4a8709e9856f9c0"} Dec 09 09:51:16 crc kubenswrapper[4824]: I1209 09:51:16.768960 4824 generic.go:334] "Generic (PLEG): container finished" podID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" containerID="63a162bdf54c6fb6ccdd3ae077bdd259528617c1ff11403d97fe992564eae348" exitCode=0 Dec 09 09:51:16 crc kubenswrapper[4824]: I1209 09:51:16.768995 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jm5ld" event={"ID":"bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271","Type":"ContainerDied","Data":"63a162bdf54c6fb6ccdd3ae077bdd259528617c1ff11403d97fe992564eae348"} Dec 09 09:51:16 crc kubenswrapper[4824]: I1209 09:51:16.769045 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jm5ld" event={"ID":"bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271","Type":"ContainerStarted","Data":"a4632c8846e6d2596d390713d5f8907592d634e3f47a6c2c002e6b3fe035806c"} Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.562444 4824 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.564030 4824 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.564287 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.564598 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39" gracePeriod=15 Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.564706 4824 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 09 09:51:18 crc kubenswrapper[4824]: E1209 09:51:18.564853 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.564872 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 09 09:51:18 crc kubenswrapper[4824]: E1209 09:51:18.564882 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.564888 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 09 09:51:18 crc kubenswrapper[4824]: E1209 09:51:18.564897 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.564903 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 09 09:51:18 crc kubenswrapper[4824]: E1209 09:51:18.564913 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.564919 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 09 09:51:18 crc kubenswrapper[4824]: E1209 09:51:18.564928 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.564933 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 09 09:51:18 crc kubenswrapper[4824]: E1209 09:51:18.564943 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.564948 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 09 09:51:18 crc kubenswrapper[4824]: E1209 09:51:18.564956 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.564961 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.565156 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.565168 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.565178 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.565185 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.565194 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.565201 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.568076 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598" gracePeriod=15 Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.568089 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90" gracePeriod=15 Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.568135 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c" gracePeriod=15 Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.568306 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8" gracePeriod=15 Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.579923 4824 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.652044 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.702187 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.702305 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.702460 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.702644 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.702735 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.702806 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.702844 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.702942 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.796454 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cbmd" event={"ID":"257d6d77-75f1-41e0-8bd2-c6ec3c1ff582","Type":"ContainerStarted","Data":"7e8ebf423a0fa29f0cea243a74412176b4ef9d93dc58478d0e66ed3494483962"} Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.797630 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.798388 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:18 crc kubenswrapper[4824]: E1209 09:51:18.799283 4824 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.128:6443: connect: connection refused" event="&Event{ObjectMeta:{certified-operators-2cbmd.187f833d6d10a6a5 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:certified-operators-2cbmd,UID:257d6d77-75f1-41e0-8bd2-c6ec3c1ff582,APIVersion:v1,ResourceVersion:29478,FieldPath:spec.containers{registry-server},},Reason:Pulling,Message:Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\",Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-09 09:51:18.798575269 +0000 UTC m=+235.133079936,LastTimestamp:2025-12-09 09:51:18.798575269 +0000 UTC m=+235.133079936,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.802228 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-95jfr" event={"ID":"6ffd99b7-9bb8-479e-a1a1-7f31ee006e38","Type":"ContainerStarted","Data":"1cf855518b8661ca5358ae1fd9e5a18c8e9fddda7448fb2fd36686a961879490"} Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.803282 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.803691 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.803734 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.803758 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.803796 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.803815 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.803842 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.803873 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.803897 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.803964 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.804629 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.804723 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.804799 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.804842 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.804895 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.804935 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.804977 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.805019 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.805054 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.807002 4824 generic.go:334] "Generic (PLEG): container finished" podID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" containerID="667c046e16c42d2c212a64e4555a99be3155a0ebad6f085a9d4d70551cafb8d6" exitCode=0 Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.807062 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jm5ld" event={"ID":"bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271","Type":"ContainerDied","Data":"667c046e16c42d2c212a64e4555a99be3155a0ebad6f085a9d4d70551cafb8d6"} Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.807926 4824 status_manager.go:851] "Failed to get status for pod" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" pod="openshift-marketplace/community-operators-jm5ld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jm5ld\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.808090 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.808248 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.808410 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.817376 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.822508 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.824147 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598" exitCode=0 Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.824190 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90" exitCode=0 Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.824200 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c" exitCode=0 Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.824209 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8" exitCode=2 Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.824339 4824 scope.go:117] "RemoveContainer" containerID="723ce948ae3fad37545adf0637d1271434edd3cef9f11cdd9eb4696961436042" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.828506 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-582xj" event={"ID":"f2572cf3-89f1-49d7-9246-9350703e9764","Type":"ContainerStarted","Data":"b70d423ba00ad8d15b6e60bc0af43894068b4f8a5957df3d2fa34df475dfb2cf"} Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.830234 4824 status_manager.go:851] "Failed to get status for pod" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" pod="openshift-marketplace/redhat-operators-582xj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-582xj\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.830564 4824 status_manager.go:851] "Failed to get status for pod" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" pod="openshift-marketplace/community-operators-jm5ld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jm5ld\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.831055 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.833499 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.834229 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:18 crc kubenswrapper[4824]: I1209 09:51:18.949582 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 09:51:18 crc kubenswrapper[4824]: W1209 09:51:18.974824 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-7026c30292f185db32b3d5532f5ff4741a40c647799cae72888b2a4046e21559 WatchSource:0}: Error finding container 7026c30292f185db32b3d5532f5ff4741a40c647799cae72888b2a4046e21559: Status 404 returned error can't find the container with id 7026c30292f185db32b3d5532f5ff4741a40c647799cae72888b2a4046e21559 Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.841724 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jm5ld" event={"ID":"bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271","Type":"ContainerStarted","Data":"4837611c2a862dc0e89e9c677ca72373bccca69d09dbaaa35de3699394b0a701"} Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.842699 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.842972 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.843226 4824 status_manager.go:851] "Failed to get status for pod" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" pod="openshift-marketplace/redhat-operators-582xj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-582xj\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.843408 4824 status_manager.go:851] "Failed to get status for pod" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" pod="openshift-marketplace/community-operators-jm5ld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jm5ld\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.843613 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.846102 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.848053 4824 generic.go:334] "Generic (PLEG): container finished" podID="332c50f5-f531-4dc6-9f0f-9d903013feff" containerID="956fcedd182c664780197779f6d8b8c482bd461e33d1c2f0145e5b98a8b3e053" exitCode=0 Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.848121 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"332c50f5-f531-4dc6-9f0f-9d903013feff","Type":"ContainerDied","Data":"956fcedd182c664780197779f6d8b8c482bd461e33d1c2f0145e5b98a8b3e053"} Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.848746 4824 status_manager.go:851] "Failed to get status for pod" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" pod="openshift-marketplace/redhat-operators-582xj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-582xj\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.849036 4824 status_manager.go:851] "Failed to get status for pod" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" pod="openshift-marketplace/community-operators-jm5ld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jm5ld\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.849288 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.849574 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.850012 4824 status_manager.go:851] "Failed to get status for pod" podUID="332c50f5-f531-4dc6-9f0f-9d903013feff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.850036 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"518105dd937b03952cbeccb9d4ed799027851c59f49636fa5a7e96cc5eeae762"} Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.850064 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"7026c30292f185db32b3d5532f5ff4741a40c647799cae72888b2a4046e21559"} Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.850504 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.850743 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.851036 4824 status_manager.go:851] "Failed to get status for pod" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" pod="openshift-marketplace/community-operators-jm5ld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jm5ld\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.851262 4824 status_manager.go:851] "Failed to get status for pod" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" pod="openshift-marketplace/redhat-operators-582xj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-582xj\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.851482 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.851770 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.852051 4824 status_manager.go:851] "Failed to get status for pod" podUID="332c50f5-f531-4dc6-9f0f-9d903013feff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.852156 4824 generic.go:334] "Generic (PLEG): container finished" podID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" containerID="7e8ebf423a0fa29f0cea243a74412176b4ef9d93dc58478d0e66ed3494483962" exitCode=0 Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.852239 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cbmd" event={"ID":"257d6d77-75f1-41e0-8bd2-c6ec3c1ff582","Type":"ContainerDied","Data":"7e8ebf423a0fa29f0cea243a74412176b4ef9d93dc58478d0e66ed3494483962"} Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.852304 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cbmd" event={"ID":"257d6d77-75f1-41e0-8bd2-c6ec3c1ff582","Type":"ContainerStarted","Data":"8eb26eb28020a6b331e841e6bc1a33e26e08169614eddc4b7de37f69670eabf9"} Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.852611 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.852922 4824 status_manager.go:851] "Failed to get status for pod" podUID="332c50f5-f531-4dc6-9f0f-9d903013feff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.853170 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.853360 4824 status_manager.go:851] "Failed to get status for pod" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" pod="openshift-marketplace/redhat-operators-582xj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-582xj\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.853540 4824 status_manager.go:851] "Failed to get status for pod" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" pod="openshift-marketplace/community-operators-jm5ld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jm5ld\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:19 crc kubenswrapper[4824]: I1209 09:51:19.853877 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.077156 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.081542 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.082680 4824 status_manager.go:851] "Failed to get status for pod" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" pod="openshift-marketplace/redhat-operators-582xj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-582xj\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.083366 4824 status_manager.go:851] "Failed to get status for pod" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" pod="openshift-marketplace/community-operators-jm5ld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jm5ld\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.083680 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.083990 4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.084709 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.085016 4824 status_manager.go:851] "Failed to get status for pod" podUID="332c50f5-f531-4dc6-9f0f-9d903013feff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.085272 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.184112 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.185104 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.185482 4824 status_manager.go:851] "Failed to get status for pod" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" pod="openshift-marketplace/redhat-operators-582xj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-582xj\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.186005 4824 status_manager.go:851] "Failed to get status for pod" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" pod="openshift-marketplace/community-operators-jm5ld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jm5ld\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.186246 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.186498 4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.186711 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.186897 4824 status_manager.go:851] "Failed to get status for pod" podUID="332c50f5-f531-4dc6-9f0f-9d903013feff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.240634 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.240937 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.240999 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.241225 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.241326 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.241352 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.241739 4824 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.241772 4824 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.241801 4824 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.343329 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/332c50f5-f531-4dc6-9f0f-9d903013feff-kubelet-dir\") pod \"332c50f5-f531-4dc6-9f0f-9d903013feff\" (UID: \"332c50f5-f531-4dc6-9f0f-9d903013feff\") " Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.343522 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/332c50f5-f531-4dc6-9f0f-9d903013feff-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "332c50f5-f531-4dc6-9f0f-9d903013feff" (UID: "332c50f5-f531-4dc6-9f0f-9d903013feff"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.343563 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/332c50f5-f531-4dc6-9f0f-9d903013feff-var-lock\") pod \"332c50f5-f531-4dc6-9f0f-9d903013feff\" (UID: \"332c50f5-f531-4dc6-9f0f-9d903013feff\") " Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.343685 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/332c50f5-f531-4dc6-9f0f-9d903013feff-var-lock" (OuterVolumeSpecName: "var-lock") pod "332c50f5-f531-4dc6-9f0f-9d903013feff" (UID: "332c50f5-f531-4dc6-9f0f-9d903013feff"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.343820 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/332c50f5-f531-4dc6-9f0f-9d903013feff-kube-api-access\") pod \"332c50f5-f531-4dc6-9f0f-9d903013feff\" (UID: \"332c50f5-f531-4dc6-9f0f-9d903013feff\") " Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.344277 4824 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/332c50f5-f531-4dc6-9f0f-9d903013feff-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.344374 4824 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/332c50f5-f531-4dc6-9f0f-9d903013feff-var-lock\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.356991 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/332c50f5-f531-4dc6-9f0f-9d903013feff-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "332c50f5-f531-4dc6-9f0f-9d903013feff" (UID: "332c50f5-f531-4dc6-9f0f-9d903013feff"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.445590 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/332c50f5-f531-4dc6-9f0f-9d903013feff-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.876525 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.877771 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"332c50f5-f531-4dc6-9f0f-9d903013feff","Type":"ContainerDied","Data":"d2de8967b2631ed5500dcb843ff5db9cf56e3b0c7c4d4e4419df468e93096ef1"} Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.877932 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2de8967b2631ed5500dcb843ff5db9cf56e3b0c7c4d4e4419df468e93096ef1" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.889460 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.891864 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39" exitCode=0 Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.891961 4824 scope.go:117] "RemoveContainer" containerID="b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.892050 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.906834 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.907061 4824 status_manager.go:851] "Failed to get status for pod" podUID="332c50f5-f531-4dc6-9f0f-9d903013feff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.907264 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.907463 4824 status_manager.go:851] "Failed to get status for pod" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" pod="openshift-marketplace/community-operators-jm5ld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jm5ld\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.907828 4824 status_manager.go:851] "Failed to get status for pod" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" pod="openshift-marketplace/redhat-operators-582xj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-582xj\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.908096 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.908371 4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.919731 4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.919964 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.920375 4824 status_manager.go:851] "Failed to get status for pod" podUID="332c50f5-f531-4dc6-9f0f-9d903013feff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.922669 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.923508 4824 status_manager.go:851] "Failed to get status for pod" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" pod="openshift-marketplace/redhat-operators-582xj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-582xj\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.924298 4824 status_manager.go:851] "Failed to get status for pod" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" pod="openshift-marketplace/community-operators-jm5ld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jm5ld\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.924936 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.925581 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.926497 4824 scope.go:117] "RemoveContainer" containerID="6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.951865 4824 scope.go:117] "RemoveContainer" containerID="5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c" Dec 09 09:51:21 crc kubenswrapper[4824]: I1209 09:51:21.988330 4824 scope.go:117] "RemoveContainer" containerID="cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8" Dec 09 09:51:22 crc kubenswrapper[4824]: I1209 09:51:22.016377 4824 scope.go:117] "RemoveContainer" containerID="7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39" Dec 09 09:51:22 crc kubenswrapper[4824]: I1209 09:51:22.048503 4824 scope.go:117] "RemoveContainer" containerID="7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b" Dec 09 09:51:22 crc kubenswrapper[4824]: I1209 09:51:22.079553 4824 scope.go:117] "RemoveContainer" containerID="b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598" Dec 09 09:51:22 crc kubenswrapper[4824]: E1209 09:51:22.080287 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\": container with ID starting with b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598 not found: ID does not exist" containerID="b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598" Dec 09 09:51:22 crc kubenswrapper[4824]: I1209 09:51:22.080324 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598"} err="failed to get container status \"b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\": rpc error: code = NotFound desc = could not find container \"b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598\": container with ID starting with b7db7af8b2d04c5e74f2689f21b04a86326b3d8dca4eff4626c7f16bbe6ae598 not found: ID does not exist" Dec 09 09:51:22 crc kubenswrapper[4824]: I1209 09:51:22.080354 4824 scope.go:117] "RemoveContainer" containerID="6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90" Dec 09 09:51:22 crc kubenswrapper[4824]: E1209 09:51:22.080961 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\": container with ID starting with 6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90 not found: ID does not exist" containerID="6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90" Dec 09 09:51:22 crc kubenswrapper[4824]: I1209 09:51:22.081003 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90"} err="failed to get container status \"6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\": rpc error: code = NotFound desc = could not find container \"6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90\": container with ID starting with 6b4d258c8f347c1ab814b23d7938fbff04bba3f62ef45bcc76907339b487ad90 not found: ID does not exist" Dec 09 09:51:22 crc kubenswrapper[4824]: I1209 09:51:22.081025 4824 scope.go:117] "RemoveContainer" containerID="5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c" Dec 09 09:51:22 crc kubenswrapper[4824]: E1209 09:51:22.081339 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\": container with ID starting with 5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c not found: ID does not exist" containerID="5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c" Dec 09 09:51:22 crc kubenswrapper[4824]: I1209 09:51:22.081366 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c"} err="failed to get container status \"5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\": rpc error: code = NotFound desc = could not find container \"5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c\": container with ID starting with 5b94cf3cd7d2a62841e907c2690777f237054d0825d75d132f7afc4479419e0c not found: ID does not exist" Dec 09 09:51:22 crc kubenswrapper[4824]: I1209 09:51:22.081381 4824 scope.go:117] "RemoveContainer" containerID="cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8" Dec 09 09:51:22 crc kubenswrapper[4824]: E1209 09:51:22.081709 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\": container with ID starting with cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8 not found: ID does not exist" containerID="cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8" Dec 09 09:51:22 crc kubenswrapper[4824]: I1209 09:51:22.081733 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8"} err="failed to get container status \"cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\": rpc error: code = NotFound desc = could not find container \"cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8\": container with ID starting with cd8bcff3a2de1d1ac68b6d714cd18877c5360d891f721c1160fa83ab20f6cab8 not found: ID does not exist" Dec 09 09:51:22 crc kubenswrapper[4824]: I1209 09:51:22.081748 4824 scope.go:117] "RemoveContainer" containerID="7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39" Dec 09 09:51:22 crc kubenswrapper[4824]: E1209 09:51:22.082134 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\": container with ID starting with 7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39 not found: ID does not exist" containerID="7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39" Dec 09 09:51:22 crc kubenswrapper[4824]: I1209 09:51:22.082157 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39"} err="failed to get container status \"7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\": rpc error: code = NotFound desc = could not find container \"7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39\": container with ID starting with 7b7cb7368205138952a3ebab7d84c76607ad4bd1f0b33414a6f67d43b37abc39 not found: ID does not exist" Dec 09 09:51:22 crc kubenswrapper[4824]: I1209 09:51:22.082172 4824 scope.go:117] "RemoveContainer" containerID="7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b" Dec 09 09:51:22 crc kubenswrapper[4824]: E1209 09:51:22.083159 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\": container with ID starting with 7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b not found: ID does not exist" containerID="7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b" Dec 09 09:51:22 crc kubenswrapper[4824]: I1209 09:51:22.083192 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b"} err="failed to get container status \"7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\": rpc error: code = NotFound desc = could not find container \"7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b\": container with ID starting with 7decca5560d6e7df080d5a4ec311bc7572236a3fe7637c0747c8cc1f3693c55b not found: ID does not exist" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.208499 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-582xj" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.208913 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-582xj" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.565962 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-95jfr" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.566014 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-95jfr" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.607106 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-95jfr" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.608046 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.608718 4824 status_manager.go:851] "Failed to get status for pod" podUID="332c50f5-f531-4dc6-9f0f-9d903013feff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.609295 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.609646 4824 status_manager.go:851] "Failed to get status for pod" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" pod="openshift-marketplace/community-operators-jm5ld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jm5ld\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.610047 4824 status_manager.go:851] "Failed to get status for pod" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" pod="openshift-marketplace/redhat-operators-582xj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-582xj\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.610388 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:23 crc kubenswrapper[4824]: E1209 09:51:23.814513 4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:23 crc kubenswrapper[4824]: E1209 09:51:23.814926 4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:23 crc kubenswrapper[4824]: E1209 09:51:23.815363 4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:23 crc kubenswrapper[4824]: E1209 09:51:23.815666 4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:23 crc kubenswrapper[4824]: E1209 09:51:23.815938 4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.815968 4824 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 09 09:51:23 crc kubenswrapper[4824]: E1209 09:51:23.816232 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.128:6443: connect: connection refused" interval="200ms" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.913618 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.914085 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.914313 4824 status_manager.go:851] "Failed to get status for pod" podUID="332c50f5-f531-4dc6-9f0f-9d903013feff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.914600 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.914896 4824 status_manager.go:851] "Failed to get status for pod" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" pod="openshift-marketplace/redhat-operators-582xj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-582xj\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.915932 4824 status_manager.go:851] "Failed to get status for pod" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" pod="openshift-marketplace/community-operators-jm5ld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jm5ld\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.969703 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-95jfr" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.972558 4824 status_manager.go:851] "Failed to get status for pod" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" pod="openshift-marketplace/redhat-operators-582xj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-582xj\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.973436 4824 status_manager.go:851] "Failed to get status for pod" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" pod="openshift-marketplace/community-operators-jm5ld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jm5ld\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.973771 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.974385 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.975540 4824 status_manager.go:851] "Failed to get status for pod" podUID="332c50f5-f531-4dc6-9f0f-9d903013feff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:23 crc kubenswrapper[4824]: I1209 09:51:23.975905 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:24 crc kubenswrapper[4824]: E1209 09:51:24.018953 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.128:6443: connect: connection refused" interval="400ms" Dec 09 09:51:24 crc kubenswrapper[4824]: I1209 09:51:24.275175 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-582xj" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" containerName="registry-server" probeResult="failure" output=< Dec 09 09:51:24 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 09:51:24 crc kubenswrapper[4824]: > Dec 09 09:51:24 crc kubenswrapper[4824]: E1209 09:51:24.420439 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.128:6443: connect: connection refused" interval="800ms" Dec 09 09:51:25 crc kubenswrapper[4824]: E1209 09:51:25.221823 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.128:6443: connect: connection refused" interval="1.6s" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.304656 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jm5ld" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.304735 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jm5ld" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.352716 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jm5ld" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.353869 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.354412 4824 status_manager.go:851] "Failed to get status for pod" podUID="332c50f5-f531-4dc6-9f0f-9d903013feff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.354772 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.355141 4824 status_manager.go:851] "Failed to get status for pod" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" pod="openshift-marketplace/redhat-operators-582xj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-582xj\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.355478 4824 status_manager.go:851] "Failed to get status for pod" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" pod="openshift-marketplace/community-operators-jm5ld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jm5ld\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.355913 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.893860 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2cbmd" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.893951 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2cbmd" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.937494 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2cbmd" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.938524 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.939480 4824 status_manager.go:851] "Failed to get status for pod" podUID="332c50f5-f531-4dc6-9f0f-9d903013feff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.940143 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.940642 4824 status_manager.go:851] "Failed to get status for pod" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" pod="openshift-marketplace/redhat-operators-582xj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-582xj\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.941185 4824 status_manager.go:851] "Failed to get status for pod" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" pod="openshift-marketplace/community-operators-jm5ld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jm5ld\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.942118 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.967140 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jm5ld" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.968134 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.969183 4824 status_manager.go:851] "Failed to get status for pod" podUID="332c50f5-f531-4dc6-9f0f-9d903013feff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.969934 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.970415 4824 status_manager.go:851] "Failed to get status for pod" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" pod="openshift-marketplace/redhat-operators-582xj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-582xj\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.971083 4824 status_manager.go:851] "Failed to get status for pod" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" pod="openshift-marketplace/community-operators-jm5ld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jm5ld\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.971420 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.987763 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2cbmd" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.988739 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.989423 4824 status_manager.go:851] "Failed to get status for pod" podUID="332c50f5-f531-4dc6-9f0f-9d903013feff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.989772 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.990146 4824 status_manager.go:851] "Failed to get status for pod" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" pod="openshift-marketplace/redhat-operators-582xj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-582xj\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.990495 4824 status_manager.go:851] "Failed to get status for pod" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" pod="openshift-marketplace/community-operators-jm5ld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jm5ld\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:25 crc kubenswrapper[4824]: I1209 09:51:25.990850 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:26 crc kubenswrapper[4824]: E1209 09:51:26.823631 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.128:6443: connect: connection refused" interval="3.2s" Dec 09 09:51:26 crc kubenswrapper[4824]: E1209 09:51:26.940093 4824 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.128:6443: connect: connection refused" event="&Event{ObjectMeta:{certified-operators-2cbmd.187f833d6d10a6a5 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:certified-operators-2cbmd,UID:257d6d77-75f1-41e0-8bd2-c6ec3c1ff582,APIVersion:v1,ResourceVersion:29478,FieldPath:spec.containers{registry-server},},Reason:Pulling,Message:Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\",Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-09 09:51:18.798575269 +0000 UTC m=+235.133079936,LastTimestamp:2025-12-09 09:51:18.798575269 +0000 UTC m=+235.133079936,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 09 09:51:30 crc kubenswrapper[4824]: E1209 09:51:30.025993 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.128:6443: connect: connection refused" interval="6.4s" Dec 09 09:51:31 crc kubenswrapper[4824]: E1209 09:51:31.792160 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:51:31Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:51:31Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:51:31Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T09:51:31Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:31 crc kubenswrapper[4824]: E1209 09:51:31.793159 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:31 crc kubenswrapper[4824]: E1209 09:51:31.793868 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:31 crc kubenswrapper[4824]: E1209 09:51:31.794168 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:31 crc kubenswrapper[4824]: E1209 09:51:31.794432 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:31 crc kubenswrapper[4824]: E1209 09:51:31.794458 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 09:51:32 crc kubenswrapper[4824]: I1209 09:51:32.968739 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 09 09:51:32 crc kubenswrapper[4824]: I1209 09:51:32.969296 4824 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0" exitCode=1 Dec 09 09:51:32 crc kubenswrapper[4824]: I1209 09:51:32.969343 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0"} Dec 09 09:51:32 crc kubenswrapper[4824]: I1209 09:51:32.970133 4824 scope.go:117] "RemoveContainer" containerID="98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0" Dec 09 09:51:32 crc kubenswrapper[4824]: I1209 09:51:32.970248 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:32 crc kubenswrapper[4824]: I1209 09:51:32.970474 4824 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:32 crc kubenswrapper[4824]: I1209 09:51:32.970853 4824 status_manager.go:851] "Failed to get status for pod" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" pod="openshift-marketplace/redhat-operators-582xj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-582xj\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:32 crc kubenswrapper[4824]: I1209 09:51:32.971432 4824 status_manager.go:851] "Failed to get status for pod" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" pod="openshift-marketplace/community-operators-jm5ld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jm5ld\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:32 crc kubenswrapper[4824]: I1209 09:51:32.971675 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:32 crc kubenswrapper[4824]: I1209 09:51:32.971892 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:32 crc kubenswrapper[4824]: I1209 09:51:32.972094 4824 status_manager.go:851] "Failed to get status for pod" podUID="332c50f5-f531-4dc6-9f0f-9d903013feff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.247104 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-582xj" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.248431 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.249128 4824 status_manager.go:851] "Failed to get status for pod" podUID="332c50f5-f531-4dc6-9f0f-9d903013feff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.249775 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.250596 4824 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.250912 4824 status_manager.go:851] "Failed to get status for pod" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" pod="openshift-marketplace/redhat-operators-582xj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-582xj\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.251466 4824 status_manager.go:851] "Failed to get status for pod" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" pod="openshift-marketplace/community-operators-jm5ld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jm5ld\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.251769 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.290230 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-582xj" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.291346 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.292290 4824 status_manager.go:851] "Failed to get status for pod" podUID="332c50f5-f531-4dc6-9f0f-9d903013feff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.293198 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.293642 4824 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.294063 4824 status_manager.go:851] "Failed to get status for pod" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" pod="openshift-marketplace/redhat-operators-582xj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-582xj\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.294400 4824 status_manager.go:851] "Failed to get status for pod" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" pod="openshift-marketplace/community-operators-jm5ld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jm5ld\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.294685 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.913080 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.913339 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.913757 4824 status_manager.go:851] "Failed to get status for pod" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" pod="openshift-marketplace/redhat-operators-582xj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-582xj\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.914118 4824 status_manager.go:851] "Failed to get status for pod" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" pod="openshift-marketplace/community-operators-jm5ld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jm5ld\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.914550 4824 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.915079 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.916214 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.916528 4824 status_manager.go:851] "Failed to get status for pod" podUID="332c50f5-f531-4dc6-9f0f-9d903013feff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.916949 4824 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.918615 4824 status_manager.go:851] "Failed to get status for pod" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" pod="openshift-marketplace/redhat-operators-582xj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-582xj\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.918941 4824 status_manager.go:851] "Failed to get status for pod" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" pod="openshift-marketplace/community-operators-jm5ld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jm5ld\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.919509 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.920029 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.920354 4824 status_manager.go:851] "Failed to get status for pod" podUID="332c50f5-f531-4dc6-9f0f-9d903013feff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.920674 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.931074 4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d9ca83da-bb34-4480-8fa1-9cca30abfb02" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.931215 4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d9ca83da-bb34-4480-8fa1-9cca30abfb02" Dec 09 09:51:33 crc kubenswrapper[4824]: E1209 09:51:33.931728 4824 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.932580 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:51:33 crc kubenswrapper[4824]: W1209 09:51:33.955810 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-6c6073b61cdb1dcbddc110259127b27d2fd1b5e9e0fcd1fba4ff8632db9327a3 WatchSource:0}: Error finding container 6c6073b61cdb1dcbddc110259127b27d2fd1b5e9e0fcd1fba4ff8632db9327a3: Status 404 returned error can't find the container with id 6c6073b61cdb1dcbddc110259127b27d2fd1b5e9e0fcd1fba4ff8632db9327a3 Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.976179 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6c6073b61cdb1dcbddc110259127b27d2fd1b5e9e0fcd1fba4ff8632db9327a3"} Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.979443 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.979637 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8b2a31f728d08d316ea20f26b82c72e791d8148fb0e65b5ccd6c881dfaf923e1"} Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.981040 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.981362 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.982027 4824 status_manager.go:851] "Failed to get status for pod" podUID="332c50f5-f531-4dc6-9f0f-9d903013feff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.982412 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.982643 4824 status_manager.go:851] "Failed to get status for pod" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" pod="openshift-marketplace/redhat-operators-582xj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-582xj\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.982904 4824 status_manager.go:851] "Failed to get status for pod" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" pod="openshift-marketplace/community-operators-jm5ld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jm5ld\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:33 crc kubenswrapper[4824]: I1209 09:51:33.983258 4824 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:34 crc kubenswrapper[4824]: I1209 09:51:34.338084 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 09:51:34 crc kubenswrapper[4824]: I1209 09:51:34.937361 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 09:51:34 crc kubenswrapper[4824]: I1209 09:51:34.937926 4824 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 09 09:51:34 crc kubenswrapper[4824]: I1209 09:51:34.938021 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 09 09:51:34 crc kubenswrapper[4824]: I1209 09:51:34.990545 4824 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="d243a4a98793d6388a75523fbbfc536590ce6ce48234599b4c2b5fd4f9bc93d7" exitCode=0 Dec 09 09:51:34 crc kubenswrapper[4824]: I1209 09:51:34.990701 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"d243a4a98793d6388a75523fbbfc536590ce6ce48234599b4c2b5fd4f9bc93d7"} Dec 09 09:51:34 crc kubenswrapper[4824]: I1209 09:51:34.990930 4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d9ca83da-bb34-4480-8fa1-9cca30abfb02" Dec 09 09:51:34 crc kubenswrapper[4824]: I1209 09:51:34.990953 4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d9ca83da-bb34-4480-8fa1-9cca30abfb02" Dec 09 09:51:34 crc kubenswrapper[4824]: E1209 09:51:34.991637 4824 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:51:34 crc kubenswrapper[4824]: I1209 09:51:34.991637 4824 status_manager.go:851] "Failed to get status for pod" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" pod="openshift-marketplace/certified-operators-2cbmd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2cbmd\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:34 crc kubenswrapper[4824]: I1209 09:51:34.992207 4824 status_manager.go:851] "Failed to get status for pod" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" pod="openshift-marketplace/redhat-marketplace-95jfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-95jfr\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:34 crc kubenswrapper[4824]: I1209 09:51:34.993026 4824 status_manager.go:851] "Failed to get status for pod" podUID="332c50f5-f531-4dc6-9f0f-9d903013feff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:34 crc kubenswrapper[4824]: I1209 09:51:34.993495 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:34 crc kubenswrapper[4824]: I1209 09:51:34.993764 4824 status_manager.go:851] "Failed to get status for pod" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" pod="openshift-marketplace/redhat-operators-582xj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-582xj\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:34 crc kubenswrapper[4824]: I1209 09:51:34.994087 4824 status_manager.go:851] "Failed to get status for pod" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" pod="openshift-marketplace/community-operators-jm5ld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jm5ld\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:34 crc kubenswrapper[4824]: I1209 09:51:34.995619 4824 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.128:6443: connect: connection refused" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:34.999992 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" podUID="a7eb77ab-add0-4601-a279-3f8fc956b52f" containerName="oauth-openshift" containerID="cri-o://baba12e8836c2703824b250129d4413aa64c3efe56d42e645c1bb5b413d11880" gracePeriod=15 Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.457085 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.515666 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-router-certs\") pod \"a7eb77ab-add0-4601-a279-3f8fc956b52f\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.515736 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-user-template-error\") pod \"a7eb77ab-add0-4601-a279-3f8fc956b52f\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.515773 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-serving-cert\") pod \"a7eb77ab-add0-4601-a279-3f8fc956b52f\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.515830 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-user-template-login\") pod \"a7eb77ab-add0-4601-a279-3f8fc956b52f\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.515863 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-cliconfig\") pod \"a7eb77ab-add0-4601-a279-3f8fc956b52f\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.515884 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a7eb77ab-add0-4601-a279-3f8fc956b52f-audit-dir\") pod \"a7eb77ab-add0-4601-a279-3f8fc956b52f\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.515914 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-user-idp-0-file-data\") pod \"a7eb77ab-add0-4601-a279-3f8fc956b52f\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.515953 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a7eb77ab-add0-4601-a279-3f8fc956b52f-audit-policies\") pod \"a7eb77ab-add0-4601-a279-3f8fc956b52f\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.515973 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2zqr\" (UniqueName: \"kubernetes.io/projected/a7eb77ab-add0-4601-a279-3f8fc956b52f-kube-api-access-v2zqr\") pod \"a7eb77ab-add0-4601-a279-3f8fc956b52f\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.516003 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-trusted-ca-bundle\") pod \"a7eb77ab-add0-4601-a279-3f8fc956b52f\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.516051 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-ocp-branding-template\") pod \"a7eb77ab-add0-4601-a279-3f8fc956b52f\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.516090 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-session\") pod \"a7eb77ab-add0-4601-a279-3f8fc956b52f\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.516117 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-service-ca\") pod \"a7eb77ab-add0-4601-a279-3f8fc956b52f\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.516161 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-user-template-provider-selection\") pod \"a7eb77ab-add0-4601-a279-3f8fc956b52f\" (UID: \"a7eb77ab-add0-4601-a279-3f8fc956b52f\") " Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.517293 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a7eb77ab-add0-4601-a279-3f8fc956b52f-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "a7eb77ab-add0-4601-a279-3f8fc956b52f" (UID: "a7eb77ab-add0-4601-a279-3f8fc956b52f"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.517761 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "a7eb77ab-add0-4601-a279-3f8fc956b52f" (UID: "a7eb77ab-add0-4601-a279-3f8fc956b52f"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.517850 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7eb77ab-add0-4601-a279-3f8fc956b52f-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "a7eb77ab-add0-4601-a279-3f8fc956b52f" (UID: "a7eb77ab-add0-4601-a279-3f8fc956b52f"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.518213 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "a7eb77ab-add0-4601-a279-3f8fc956b52f" (UID: "a7eb77ab-add0-4601-a279-3f8fc956b52f"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.518430 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "a7eb77ab-add0-4601-a279-3f8fc956b52f" (UID: "a7eb77ab-add0-4601-a279-3f8fc956b52f"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.523608 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "a7eb77ab-add0-4601-a279-3f8fc956b52f" (UID: "a7eb77ab-add0-4601-a279-3f8fc956b52f"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.523801 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "a7eb77ab-add0-4601-a279-3f8fc956b52f" (UID: "a7eb77ab-add0-4601-a279-3f8fc956b52f"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.523760 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7eb77ab-add0-4601-a279-3f8fc956b52f-kube-api-access-v2zqr" (OuterVolumeSpecName: "kube-api-access-v2zqr") pod "a7eb77ab-add0-4601-a279-3f8fc956b52f" (UID: "a7eb77ab-add0-4601-a279-3f8fc956b52f"). InnerVolumeSpecName "kube-api-access-v2zqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.524330 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "a7eb77ab-add0-4601-a279-3f8fc956b52f" (UID: "a7eb77ab-add0-4601-a279-3f8fc956b52f"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.524570 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "a7eb77ab-add0-4601-a279-3f8fc956b52f" (UID: "a7eb77ab-add0-4601-a279-3f8fc956b52f"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.525716 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "a7eb77ab-add0-4601-a279-3f8fc956b52f" (UID: "a7eb77ab-add0-4601-a279-3f8fc956b52f"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.526468 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "a7eb77ab-add0-4601-a279-3f8fc956b52f" (UID: "a7eb77ab-add0-4601-a279-3f8fc956b52f"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.528373 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "a7eb77ab-add0-4601-a279-3f8fc956b52f" (UID: "a7eb77ab-add0-4601-a279-3f8fc956b52f"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.528953 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "a7eb77ab-add0-4601-a279-3f8fc956b52f" (UID: "a7eb77ab-add0-4601-a279-3f8fc956b52f"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.623011 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2zqr\" (UniqueName: \"kubernetes.io/projected/a7eb77ab-add0-4601-a279-3f8fc956b52f-kube-api-access-v2zqr\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.626194 4824 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a7eb77ab-add0-4601-a279-3f8fc956b52f-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.626211 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.626226 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.626238 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.626249 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.626262 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.626274 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.626285 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.626295 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.626305 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.626316 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.626326 4824 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a7eb77ab-add0-4601-a279-3f8fc956b52f-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:35 crc kubenswrapper[4824]: I1209 09:51:35.626339 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a7eb77ab-add0-4601-a279-3f8fc956b52f-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 09 09:51:36 crc kubenswrapper[4824]: I1209 09:51:36.016713 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9019863b2f46d5fbf0a3f1b16ee2a5ec061f52bbde3123dea2f5d6013afb1f1a"} Dec 09 09:51:36 crc kubenswrapper[4824]: I1209 09:51:36.016801 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"76ecb206a0235ba181852f679df255b2ef89778f52e14db010f9c4df684147af"} Dec 09 09:51:36 crc kubenswrapper[4824]: I1209 09:51:36.016816 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2ffd1cac0ccbc36902aee091df11233553fa81e7d1249fd0016b7e388b8f39d8"} Dec 09 09:51:36 crc kubenswrapper[4824]: I1209 09:51:36.022920 4824 generic.go:334] "Generic (PLEG): container finished" podID="a7eb77ab-add0-4601-a279-3f8fc956b52f" containerID="baba12e8836c2703824b250129d4413aa64c3efe56d42e645c1bb5b413d11880" exitCode=0 Dec 09 09:51:36 crc kubenswrapper[4824]: I1209 09:51:36.023017 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" Dec 09 09:51:36 crc kubenswrapper[4824]: I1209 09:51:36.023074 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" event={"ID":"a7eb77ab-add0-4601-a279-3f8fc956b52f","Type":"ContainerDied","Data":"baba12e8836c2703824b250129d4413aa64c3efe56d42e645c1bb5b413d11880"} Dec 09 09:51:36 crc kubenswrapper[4824]: I1209 09:51:36.023171 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-l5ld8" event={"ID":"a7eb77ab-add0-4601-a279-3f8fc956b52f","Type":"ContainerDied","Data":"6da487162b1512cf0a7f19ec366feec1802e6cb6f9163c7b9a8b00823c0c4046"} Dec 09 09:51:36 crc kubenswrapper[4824]: I1209 09:51:36.023199 4824 scope.go:117] "RemoveContainer" containerID="baba12e8836c2703824b250129d4413aa64c3efe56d42e645c1bb5b413d11880" Dec 09 09:51:36 crc kubenswrapper[4824]: I1209 09:51:36.043843 4824 scope.go:117] "RemoveContainer" containerID="baba12e8836c2703824b250129d4413aa64c3efe56d42e645c1bb5b413d11880" Dec 09 09:51:36 crc kubenswrapper[4824]: E1209 09:51:36.044818 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"baba12e8836c2703824b250129d4413aa64c3efe56d42e645c1bb5b413d11880\": container with ID starting with baba12e8836c2703824b250129d4413aa64c3efe56d42e645c1bb5b413d11880 not found: ID does not exist" containerID="baba12e8836c2703824b250129d4413aa64c3efe56d42e645c1bb5b413d11880" Dec 09 09:51:36 crc kubenswrapper[4824]: I1209 09:51:36.044890 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baba12e8836c2703824b250129d4413aa64c3efe56d42e645c1bb5b413d11880"} err="failed to get container status \"baba12e8836c2703824b250129d4413aa64c3efe56d42e645c1bb5b413d11880\": rpc error: code = NotFound desc = could not find container \"baba12e8836c2703824b250129d4413aa64c3efe56d42e645c1bb5b413d11880\": container with ID starting with baba12e8836c2703824b250129d4413aa64c3efe56d42e645c1bb5b413d11880 not found: ID does not exist" Dec 09 09:51:37 crc kubenswrapper[4824]: I1209 09:51:37.033249 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a992ea952204f3a538829e8170c9fa071564f1db41e7412da1b3dda944298633"} Dec 09 09:51:37 crc kubenswrapper[4824]: I1209 09:51:37.033678 4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d9ca83da-bb34-4480-8fa1-9cca30abfb02" Dec 09 09:51:37 crc kubenswrapper[4824]: I1209 09:51:37.033718 4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d9ca83da-bb34-4480-8fa1-9cca30abfb02" Dec 09 09:51:37 crc kubenswrapper[4824]: I1209 09:51:37.033727 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:51:37 crc kubenswrapper[4824]: I1209 09:51:37.033746 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a872f519ecb6a2d089d25b6c2f24470a18d06a6abcf0097f164f9c2035618540"} Dec 09 09:51:38 crc kubenswrapper[4824]: I1209 09:51:38.932944 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:51:38 crc kubenswrapper[4824]: I1209 09:51:38.933284 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:51:38 crc kubenswrapper[4824]: I1209 09:51:38.938708 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:51:42 crc kubenswrapper[4824]: I1209 09:51:42.060937 4824 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:51:43 crc kubenswrapper[4824]: I1209 09:51:43.075334 4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d9ca83da-bb34-4480-8fa1-9cca30abfb02" Dec 09 09:51:43 crc kubenswrapper[4824]: I1209 09:51:43.075381 4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d9ca83da-bb34-4480-8fa1-9cca30abfb02" Dec 09 09:51:43 crc kubenswrapper[4824]: I1209 09:51:43.081556 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:51:43 crc kubenswrapper[4824]: I1209 09:51:43.929755 4824 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="19ca3449-968c-4513-9735-dd8cc0af6fc8" Dec 09 09:51:44 crc kubenswrapper[4824]: I1209 09:51:44.081664 4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d9ca83da-bb34-4480-8fa1-9cca30abfb02" Dec 09 09:51:44 crc kubenswrapper[4824]: I1209 09:51:44.081700 4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d9ca83da-bb34-4480-8fa1-9cca30abfb02" Dec 09 09:51:44 crc kubenswrapper[4824]: I1209 09:51:44.086726 4824 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="19ca3449-968c-4513-9735-dd8cc0af6fc8" Dec 09 09:51:44 crc kubenswrapper[4824]: I1209 09:51:44.938255 4824 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 09 09:51:44 crc kubenswrapper[4824]: I1209 09:51:44.938376 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 09 09:51:51 crc kubenswrapper[4824]: I1209 09:51:51.682265 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 09 09:51:51 crc kubenswrapper[4824]: I1209 09:51:51.891432 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 09 09:51:52 crc kubenswrapper[4824]: I1209 09:51:52.520254 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 09 09:51:52 crc kubenswrapper[4824]: I1209 09:51:52.535163 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 09 09:51:52 crc kubenswrapper[4824]: I1209 09:51:52.735415 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 09 09:51:52 crc kubenswrapper[4824]: I1209 09:51:52.802085 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 09 09:51:52 crc kubenswrapper[4824]: I1209 09:51:52.802141 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 09 09:51:53 crc kubenswrapper[4824]: I1209 09:51:53.286488 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 09 09:51:53 crc kubenswrapper[4824]: I1209 09:51:53.580113 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 09 09:51:53 crc kubenswrapper[4824]: I1209 09:51:53.733092 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 09 09:51:54 crc kubenswrapper[4824]: I1209 09:51:54.130951 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 09 09:51:54 crc kubenswrapper[4824]: I1209 09:51:54.215757 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 09 09:51:54 crc kubenswrapper[4824]: I1209 09:51:54.813648 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 09 09:51:54 crc kubenswrapper[4824]: I1209 09:51:54.866825 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 09 09:51:54 crc kubenswrapper[4824]: I1209 09:51:54.916540 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 09 09:51:54 crc kubenswrapper[4824]: I1209 09:51:54.937331 4824 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 09 09:51:54 crc kubenswrapper[4824]: I1209 09:51:54.937742 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 09 09:51:54 crc kubenswrapper[4824]: I1209 09:51:54.938013 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 09:51:54 crc kubenswrapper[4824]: I1209 09:51:54.938922 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"8b2a31f728d08d316ea20f26b82c72e791d8148fb0e65b5ccd6c881dfaf923e1"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Dec 09 09:51:54 crc kubenswrapper[4824]: I1209 09:51:54.939124 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://8b2a31f728d08d316ea20f26b82c72e791d8148fb0e65b5ccd6c881dfaf923e1" gracePeriod=30 Dec 09 09:51:54 crc kubenswrapper[4824]: I1209 09:51:54.950904 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 09 09:51:54 crc kubenswrapper[4824]: I1209 09:51:54.984890 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 09 09:51:55 crc kubenswrapper[4824]: I1209 09:51:55.013690 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 09 09:51:55 crc kubenswrapper[4824]: I1209 09:51:55.044511 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 09 09:51:55 crc kubenswrapper[4824]: I1209 09:51:55.119542 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 09 09:51:55 crc kubenswrapper[4824]: I1209 09:51:55.136259 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 09 09:51:55 crc kubenswrapper[4824]: I1209 09:51:55.185731 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 09 09:51:55 crc kubenswrapper[4824]: I1209 09:51:55.400495 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 09 09:51:55 crc kubenswrapper[4824]: I1209 09:51:55.505573 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 09 09:51:55 crc kubenswrapper[4824]: I1209 09:51:55.586999 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 09 09:51:55 crc kubenswrapper[4824]: I1209 09:51:55.588451 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 09 09:51:55 crc kubenswrapper[4824]: I1209 09:51:55.614918 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 09 09:51:55 crc kubenswrapper[4824]: I1209 09:51:55.751068 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 09 09:51:55 crc kubenswrapper[4824]: I1209 09:51:55.768914 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 09 09:51:55 crc kubenswrapper[4824]: I1209 09:51:55.770099 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 09 09:51:55 crc kubenswrapper[4824]: I1209 09:51:55.816372 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 09 09:51:55 crc kubenswrapper[4824]: I1209 09:51:55.905234 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 09 09:51:55 crc kubenswrapper[4824]: I1209 09:51:55.913979 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 09 09:51:55 crc kubenswrapper[4824]: I1209 09:51:55.915362 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 09 09:51:56 crc kubenswrapper[4824]: I1209 09:51:56.130304 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 09 09:51:56 crc kubenswrapper[4824]: I1209 09:51:56.151624 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 09 09:51:56 crc kubenswrapper[4824]: I1209 09:51:56.435543 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 09 09:51:56 crc kubenswrapper[4824]: I1209 09:51:56.451856 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 09 09:51:56 crc kubenswrapper[4824]: I1209 09:51:56.475204 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 09 09:51:56 crc kubenswrapper[4824]: I1209 09:51:56.626680 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 09 09:51:56 crc kubenswrapper[4824]: I1209 09:51:56.626773 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 09 09:51:56 crc kubenswrapper[4824]: I1209 09:51:56.633490 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 09 09:51:56 crc kubenswrapper[4824]: I1209 09:51:56.664110 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 09 09:51:56 crc kubenswrapper[4824]: I1209 09:51:56.687851 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 09 09:51:56 crc kubenswrapper[4824]: I1209 09:51:56.699565 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 09 09:51:56 crc kubenswrapper[4824]: I1209 09:51:56.756324 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 09 09:51:56 crc kubenswrapper[4824]: I1209 09:51:56.811291 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 09 09:51:56 crc kubenswrapper[4824]: I1209 09:51:56.896474 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 09 09:51:56 crc kubenswrapper[4824]: I1209 09:51:56.903747 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 09 09:51:56 crc kubenswrapper[4824]: I1209 09:51:56.951171 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 09 09:51:56 crc kubenswrapper[4824]: I1209 09:51:56.960999 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 09 09:51:57 crc kubenswrapper[4824]: I1209 09:51:57.016060 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 09 09:51:57 crc kubenswrapper[4824]: I1209 09:51:57.034751 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 09 09:51:57 crc kubenswrapper[4824]: I1209 09:51:57.091814 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 09 09:51:57 crc kubenswrapper[4824]: I1209 09:51:57.356299 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 09 09:51:57 crc kubenswrapper[4824]: I1209 09:51:57.456042 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 09 09:51:57 crc kubenswrapper[4824]: I1209 09:51:57.536043 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 09 09:51:57 crc kubenswrapper[4824]: I1209 09:51:57.662866 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 09 09:51:57 crc kubenswrapper[4824]: I1209 09:51:57.672918 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 09 09:51:57 crc kubenswrapper[4824]: I1209 09:51:57.681309 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 09 09:51:57 crc kubenswrapper[4824]: I1209 09:51:57.730019 4824 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 09 09:51:57 crc kubenswrapper[4824]: I1209 09:51:57.741381 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 09 09:51:57 crc kubenswrapper[4824]: I1209 09:51:57.747491 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 09 09:51:57 crc kubenswrapper[4824]: I1209 09:51:57.830074 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 09 09:51:57 crc kubenswrapper[4824]: I1209 09:51:57.865105 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 09 09:51:57 crc kubenswrapper[4824]: I1209 09:51:57.883392 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 09 09:51:58 crc kubenswrapper[4824]: I1209 09:51:58.000664 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 09 09:51:58 crc kubenswrapper[4824]: I1209 09:51:58.102293 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 09 09:51:58 crc kubenswrapper[4824]: I1209 09:51:58.350567 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 09 09:51:58 crc kubenswrapper[4824]: I1209 09:51:58.396858 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 09 09:51:58 crc kubenswrapper[4824]: I1209 09:51:58.433835 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 09 09:51:58 crc kubenswrapper[4824]: I1209 09:51:58.451654 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 09 09:51:58 crc kubenswrapper[4824]: I1209 09:51:58.457598 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 09 09:51:58 crc kubenswrapper[4824]: I1209 09:51:58.525934 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 09 09:51:58 crc kubenswrapper[4824]: I1209 09:51:58.577797 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 09 09:51:58 crc kubenswrapper[4824]: I1209 09:51:58.607364 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 09 09:51:58 crc kubenswrapper[4824]: I1209 09:51:58.621378 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 09 09:51:58 crc kubenswrapper[4824]: I1209 09:51:58.767360 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 09 09:51:58 crc kubenswrapper[4824]: I1209 09:51:58.833951 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 09 09:51:58 crc kubenswrapper[4824]: I1209 09:51:58.841978 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 09 09:51:58 crc kubenswrapper[4824]: I1209 09:51:58.848088 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 09 09:51:58 crc kubenswrapper[4824]: I1209 09:51:58.864471 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 09 09:51:58 crc kubenswrapper[4824]: I1209 09:51:58.954479 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 09 09:51:59 crc kubenswrapper[4824]: I1209 09:51:59.123212 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 09 09:51:59 crc kubenswrapper[4824]: I1209 09:51:59.141371 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 09 09:51:59 crc kubenswrapper[4824]: I1209 09:51:59.158633 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 09 09:51:59 crc kubenswrapper[4824]: I1209 09:51:59.159724 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 09 09:51:59 crc kubenswrapper[4824]: I1209 09:51:59.189170 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 09 09:51:59 crc kubenswrapper[4824]: I1209 09:51:59.190065 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 09 09:51:59 crc kubenswrapper[4824]: I1209 09:51:59.218456 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 09 09:51:59 crc kubenswrapper[4824]: I1209 09:51:59.245664 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 09 09:51:59 crc kubenswrapper[4824]: I1209 09:51:59.336178 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 09 09:51:59 crc kubenswrapper[4824]: I1209 09:51:59.368510 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 09 09:51:59 crc kubenswrapper[4824]: I1209 09:51:59.428027 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 09 09:51:59 crc kubenswrapper[4824]: I1209 09:51:59.514184 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 09 09:51:59 crc kubenswrapper[4824]: I1209 09:51:59.521401 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 09 09:51:59 crc kubenswrapper[4824]: I1209 09:51:59.572831 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 09 09:51:59 crc kubenswrapper[4824]: I1209 09:51:59.615011 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 09 09:51:59 crc kubenswrapper[4824]: I1209 09:51:59.723904 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 09 09:51:59 crc kubenswrapper[4824]: I1209 09:51:59.847100 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 09 09:51:59 crc kubenswrapper[4824]: I1209 09:51:59.932122 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 09 09:52:00 crc kubenswrapper[4824]: I1209 09:52:00.075806 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 09 09:52:00 crc kubenswrapper[4824]: I1209 09:52:00.122709 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 09 09:52:00 crc kubenswrapper[4824]: I1209 09:52:00.245580 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 09 09:52:00 crc kubenswrapper[4824]: I1209 09:52:00.245683 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 09 09:52:00 crc kubenswrapper[4824]: I1209 09:52:00.268578 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 09 09:52:00 crc kubenswrapper[4824]: I1209 09:52:00.309244 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 09 09:52:00 crc kubenswrapper[4824]: I1209 09:52:00.324859 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 09 09:52:00 crc kubenswrapper[4824]: I1209 09:52:00.350077 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 09 09:52:00 crc kubenswrapper[4824]: I1209 09:52:00.508731 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 09 09:52:00 crc kubenswrapper[4824]: I1209 09:52:00.531733 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 09 09:52:00 crc kubenswrapper[4824]: I1209 09:52:00.552187 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 09 09:52:00 crc kubenswrapper[4824]: I1209 09:52:00.643985 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 09 09:52:00 crc kubenswrapper[4824]: I1209 09:52:00.690386 4824 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 09 09:52:00 crc kubenswrapper[4824]: I1209 09:52:00.721490 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 09 09:52:00 crc kubenswrapper[4824]: I1209 09:52:00.812354 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 09 09:52:00 crc kubenswrapper[4824]: I1209 09:52:00.812354 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 09 09:52:00 crc kubenswrapper[4824]: I1209 09:52:00.838111 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 09 09:52:00 crc kubenswrapper[4824]: I1209 09:52:00.952709 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 09 09:52:00 crc kubenswrapper[4824]: I1209 09:52:00.974102 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 09 09:52:00 crc kubenswrapper[4824]: I1209 09:52:00.999692 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 09 09:52:01 crc kubenswrapper[4824]: I1209 09:52:01.008244 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 09 09:52:01 crc kubenswrapper[4824]: I1209 09:52:01.043534 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 09 09:52:01 crc kubenswrapper[4824]: I1209 09:52:01.066450 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 09 09:52:01 crc kubenswrapper[4824]: I1209 09:52:01.071221 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 09 09:52:01 crc kubenswrapper[4824]: I1209 09:52:01.081628 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 09 09:52:01 crc kubenswrapper[4824]: I1209 09:52:01.241202 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 09 09:52:01 crc kubenswrapper[4824]: I1209 09:52:01.248000 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 09 09:52:01 crc kubenswrapper[4824]: I1209 09:52:01.516229 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 09 09:52:01 crc kubenswrapper[4824]: I1209 09:52:01.533905 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 09 09:52:01 crc kubenswrapper[4824]: I1209 09:52:01.572961 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 09 09:52:01 crc kubenswrapper[4824]: I1209 09:52:01.635428 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 09 09:52:01 crc kubenswrapper[4824]: I1209 09:52:01.765312 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 09 09:52:01 crc kubenswrapper[4824]: I1209 09:52:01.770323 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 09 09:52:01 crc kubenswrapper[4824]: I1209 09:52:01.804352 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 09 09:52:01 crc kubenswrapper[4824]: I1209 09:52:01.864076 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 09 09:52:01 crc kubenswrapper[4824]: I1209 09:52:01.864090 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 09 09:52:01 crc kubenswrapper[4824]: I1209 09:52:01.913180 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 09 09:52:01 crc kubenswrapper[4824]: I1209 09:52:01.926419 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 09 09:52:01 crc kubenswrapper[4824]: I1209 09:52:01.951974 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 09 09:52:02 crc kubenswrapper[4824]: I1209 09:52:02.057627 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 09 09:52:02 crc kubenswrapper[4824]: I1209 09:52:02.082915 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 09 09:52:02 crc kubenswrapper[4824]: I1209 09:52:02.214925 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 09 09:52:02 crc kubenswrapper[4824]: I1209 09:52:02.413107 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 09 09:52:02 crc kubenswrapper[4824]: I1209 09:52:02.436986 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 09 09:52:02 crc kubenswrapper[4824]: I1209 09:52:02.467068 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 09 09:52:02 crc kubenswrapper[4824]: I1209 09:52:02.516456 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 09 09:52:02 crc kubenswrapper[4824]: I1209 09:52:02.551768 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 09 09:52:02 crc kubenswrapper[4824]: I1209 09:52:02.557125 4824 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 09 09:52:02 crc kubenswrapper[4824]: I1209 09:52:02.589647 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 09 09:52:02 crc kubenswrapper[4824]: I1209 09:52:02.664987 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 09 09:52:02 crc kubenswrapper[4824]: I1209 09:52:02.798832 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 09 09:52:02 crc kubenswrapper[4824]: I1209 09:52:02.952122 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 09 09:52:02 crc kubenswrapper[4824]: I1209 09:52:02.968000 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 09 09:52:02 crc kubenswrapper[4824]: I1209 09:52:02.972914 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.078747 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.079181 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.195878 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.201337 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.319174 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.414167 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.450452 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.455179 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.483968 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.513067 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.605608 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.693689 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.905276 4824 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.905865 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2cbmd" podStartSLOduration=46.357743666 podStartE2EDuration="48.905832965s" podCreationTimestamp="2025-12-09 09:51:15 +0000 UTC" firstStartedPulling="2025-12-09 09:51:16.763928231 +0000 UTC m=+233.098432898" lastFinishedPulling="2025-12-09 09:51:19.31201753 +0000 UTC m=+235.646522197" observedRunningTime="2025-12-09 09:51:41.999023587 +0000 UTC m=+258.333528254" watchObservedRunningTime="2025-12-09 09:52:03.905832965 +0000 UTC m=+280.240337632" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.908546 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-95jfr" podStartSLOduration=47.899391044 podStartE2EDuration="50.90853389s" podCreationTimestamp="2025-12-09 09:51:13 +0000 UTC" firstStartedPulling="2025-12-09 09:51:14.730841794 +0000 UTC m=+231.065346461" lastFinishedPulling="2025-12-09 09:51:17.73998464 +0000 UTC m=+234.074489307" observedRunningTime="2025-12-09 09:51:41.830961804 +0000 UTC m=+258.165466491" watchObservedRunningTime="2025-12-09 09:52:03.90853389 +0000 UTC m=+280.243038547" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.909847 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=45.909841352 podStartE2EDuration="45.909841352s" podCreationTimestamp="2025-12-09 09:51:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:51:41.901624745 +0000 UTC m=+258.236129412" watchObservedRunningTime="2025-12-09 09:52:03.909841352 +0000 UTC m=+280.244346019" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.909920 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-582xj" podStartSLOduration=48.850644978 podStartE2EDuration="51.909913014s" podCreationTimestamp="2025-12-09 09:51:12 +0000 UTC" firstStartedPulling="2025-12-09 09:51:14.735675863 +0000 UTC m=+231.070180530" lastFinishedPulling="2025-12-09 09:51:17.794943899 +0000 UTC m=+234.129448566" observedRunningTime="2025-12-09 09:51:41.956123691 +0000 UTC m=+258.290628358" watchObservedRunningTime="2025-12-09 09:52:03.909913014 +0000 UTC m=+280.244417691" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.910192 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jm5ld" podStartSLOduration=47.413474335 podStartE2EDuration="49.910186863s" podCreationTimestamp="2025-12-09 09:51:14 +0000 UTC" firstStartedPulling="2025-12-09 09:51:16.770554078 +0000 UTC m=+233.105058745" lastFinishedPulling="2025-12-09 09:51:19.267266606 +0000 UTC m=+235.601771273" observedRunningTime="2025-12-09 09:51:41.974833707 +0000 UTC m=+258.309338374" watchObservedRunningTime="2025-12-09 09:52:03.910186863 +0000 UTC m=+280.244691520" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.921612 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-l5ld8","openshift-kube-apiserver/kube-apiserver-crc"] Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.921821 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-f578d5c8f-5jzzv","openshift-kube-apiserver/kube-apiserver-crc"] Dec 09 09:52:03 crc kubenswrapper[4824]: E1209 09:52:03.922096 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="332c50f5-f531-4dc6-9f0f-9d903013feff" containerName="installer" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.922137 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="332c50f5-f531-4dc6-9f0f-9d903013feff" containerName="installer" Dec 09 09:52:03 crc kubenswrapper[4824]: E1209 09:52:03.922151 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7eb77ab-add0-4601-a279-3f8fc956b52f" containerName="oauth-openshift" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.922162 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7eb77ab-add0-4601-a279-3f8fc956b52f" containerName="oauth-openshift" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.922142 4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d9ca83da-bb34-4480-8fa1-9cca30abfb02" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.922265 4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d9ca83da-bb34-4480-8fa1-9cca30abfb02" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.922479 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="332c50f5-f531-4dc6-9f0f-9d903013feff" containerName="installer" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.922507 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7eb77ab-add0-4601-a279-3f8fc956b52f" containerName="oauth-openshift" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.923301 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.926995 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.927025 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.927253 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.927287 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.927437 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.926995 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.927293 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.927357 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.927734 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.927892 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.928064 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.928147 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.928273 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.936346 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.939792 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.948510 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.950889 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.995408 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 09 09:52:03 crc kubenswrapper[4824]: I1209 09:52:03.997609 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=21.99759101 podStartE2EDuration="21.99759101s" podCreationTimestamp="2025-12-09 09:51:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:52:03.995826415 +0000 UTC m=+280.330331112" watchObservedRunningTime="2025-12-09 09:52:03.99759101 +0000 UTC m=+280.332095677" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.012482 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.057596 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.057679 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-audit-dir\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.058087 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-system-service-ca\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.058152 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-system-serving-cert\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.058227 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbcrh\" (UniqueName: \"kubernetes.io/projected/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-kube-api-access-sbcrh\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.058278 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-user-template-error\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.058335 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.058365 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-system-router-certs\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.058729 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-system-session\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.058824 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-user-template-login\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.058921 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.058953 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-system-cliconfig\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.058984 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.059021 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-audit-policies\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.094385 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.098069 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.160512 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-user-template-error\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.160591 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.160629 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-system-router-certs\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.160698 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-system-session\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.160736 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-user-template-login\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.160799 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.160831 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-system-cliconfig\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.162177 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.162269 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-audit-policies\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.162358 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.162395 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-audit-dir\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.162474 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-system-service-ca\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.162508 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-system-serving-cert\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.162548 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbcrh\" (UniqueName: \"kubernetes.io/projected/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-kube-api-access-sbcrh\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.163151 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-audit-policies\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.163346 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-audit-dir\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.163460 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-system-service-ca\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.163691 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-system-cliconfig\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.164203 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.168318 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.169888 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-system-router-certs\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.170175 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.170507 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.170954 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-user-template-login\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.172168 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-system-serving-cert\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.174641 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-system-session\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.180813 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-v4-0-config-user-template-error\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.184249 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbcrh\" (UniqueName: \"kubernetes.io/projected/3d1e2cd1-5fc5-4a76-b1ff-07db11b07708-kube-api-access-sbcrh\") pod \"oauth-openshift-f578d5c8f-5jzzv\" (UID: \"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708\") " pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.256749 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.293473 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.350893 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.378267 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.394284 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.487373 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.563625 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.592010 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.629337 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.653772 4824 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.654318 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://518105dd937b03952cbeccb9d4ed799027851c59f49636fa5a7e96cc5eeae762" gracePeriod=5 Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.684499 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.704415 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.705822 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-f578d5c8f-5jzzv"] Dec 09 09:52:04 crc kubenswrapper[4824]: W1209 09:52:04.715210 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d1e2cd1_5fc5_4a76_b1ff_07db11b07708.slice/crio-0d751bb8b4dc74c4e48ba4db94d3909a9a831f6dc1dfa1e6feb546ea9f717ac8 WatchSource:0}: Error finding container 0d751bb8b4dc74c4e48ba4db94d3909a9a831f6dc1dfa1e6feb546ea9f717ac8: Status 404 returned error can't find the container with id 0d751bb8b4dc74c4e48ba4db94d3909a9a831f6dc1dfa1e6feb546ea9f717ac8 Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.775776 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.791220 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.803754 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.804847 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 09 09:52:04 crc kubenswrapper[4824]: I1209 09:52:04.834156 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 09 09:52:05 crc kubenswrapper[4824]: I1209 09:52:05.102603 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 09 09:52:05 crc kubenswrapper[4824]: I1209 09:52:05.110419 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 09 09:52:05 crc kubenswrapper[4824]: I1209 09:52:05.158082 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 09 09:52:05 crc kubenswrapper[4824]: I1209 09:52:05.188380 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 09 09:52:05 crc kubenswrapper[4824]: I1209 09:52:05.211317 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" event={"ID":"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708","Type":"ContainerStarted","Data":"166928d0068d6f77eb364346c00c3dd78dbb7602e7badaa6af86f3ce7f5c0df1"} Dec 09 09:52:05 crc kubenswrapper[4824]: I1209 09:52:05.211376 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" event={"ID":"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708","Type":"ContainerStarted","Data":"0d751bb8b4dc74c4e48ba4db94d3909a9a831f6dc1dfa1e6feb546ea9f717ac8"} Dec 09 09:52:05 crc kubenswrapper[4824]: I1209 09:52:05.212821 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:05 crc kubenswrapper[4824]: I1209 09:52:05.228972 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 09 09:52:05 crc kubenswrapper[4824]: I1209 09:52:05.288379 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 09 09:52:05 crc kubenswrapper[4824]: I1209 09:52:05.303426 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 09 09:52:05 crc kubenswrapper[4824]: I1209 09:52:05.515829 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 09 09:52:05 crc kubenswrapper[4824]: I1209 09:52:05.529983 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 09:52:05 crc kubenswrapper[4824]: I1209 09:52:05.559057 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" podStartSLOduration=56.559032268 podStartE2EDuration="56.559032268s" podCreationTimestamp="2025-12-09 09:51:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:52:05.242978382 +0000 UTC m=+281.577483059" watchObservedRunningTime="2025-12-09 09:52:05.559032268 +0000 UTC m=+281.893536945" Dec 09 09:52:05 crc kubenswrapper[4824]: I1209 09:52:05.585888 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 09 09:52:05 crc kubenswrapper[4824]: I1209 09:52:05.643397 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 09 09:52:05 crc kubenswrapper[4824]: I1209 09:52:05.928893 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7eb77ab-add0-4601-a279-3f8fc956b52f" path="/var/lib/kubelet/pods/a7eb77ab-add0-4601-a279-3f8fc956b52f/volumes" Dec 09 09:52:06 crc kubenswrapper[4824]: I1209 09:52:06.112164 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 09 09:52:06 crc kubenswrapper[4824]: I1209 09:52:06.125001 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 09 09:52:06 crc kubenswrapper[4824]: I1209 09:52:06.166910 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 09 09:52:06 crc kubenswrapper[4824]: I1209 09:52:06.407053 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 09 09:52:06 crc kubenswrapper[4824]: I1209 09:52:06.416396 4824 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 09 09:52:06 crc kubenswrapper[4824]: I1209 09:52:06.418624 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 09 09:52:06 crc kubenswrapper[4824]: I1209 09:52:06.659490 4824 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 09 09:52:06 crc kubenswrapper[4824]: I1209 09:52:06.812165 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 09 09:52:07 crc kubenswrapper[4824]: I1209 09:52:07.005591 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 09 09:52:07 crc kubenswrapper[4824]: I1209 09:52:07.042048 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 09 09:52:07 crc kubenswrapper[4824]: I1209 09:52:07.147274 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 09 09:52:07 crc kubenswrapper[4824]: I1209 09:52:07.247459 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 09 09:52:07 crc kubenswrapper[4824]: I1209 09:52:07.276448 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 09 09:52:07 crc kubenswrapper[4824]: I1209 09:52:07.455058 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 09 09:52:07 crc kubenswrapper[4824]: I1209 09:52:07.596684 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 09 09:52:07 crc kubenswrapper[4824]: I1209 09:52:07.652568 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 09 09:52:10 crc kubenswrapper[4824]: I1209 09:52:10.226956 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 09 09:52:10 crc kubenswrapper[4824]: I1209 09:52:10.227470 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 09:52:10 crc kubenswrapper[4824]: I1209 09:52:10.244413 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 09 09:52:10 crc kubenswrapper[4824]: I1209 09:52:10.244487 4824 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="518105dd937b03952cbeccb9d4ed799027851c59f49636fa5a7e96cc5eeae762" exitCode=137 Dec 09 09:52:10 crc kubenswrapper[4824]: I1209 09:52:10.244563 4824 scope.go:117] "RemoveContainer" containerID="518105dd937b03952cbeccb9d4ed799027851c59f49636fa5a7e96cc5eeae762" Dec 09 09:52:10 crc kubenswrapper[4824]: I1209 09:52:10.244592 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 09:52:10 crc kubenswrapper[4824]: I1209 09:52:10.264734 4824 scope.go:117] "RemoveContainer" containerID="518105dd937b03952cbeccb9d4ed799027851c59f49636fa5a7e96cc5eeae762" Dec 09 09:52:10 crc kubenswrapper[4824]: E1209 09:52:10.265619 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"518105dd937b03952cbeccb9d4ed799027851c59f49636fa5a7e96cc5eeae762\": container with ID starting with 518105dd937b03952cbeccb9d4ed799027851c59f49636fa5a7e96cc5eeae762 not found: ID does not exist" containerID="518105dd937b03952cbeccb9d4ed799027851c59f49636fa5a7e96cc5eeae762" Dec 09 09:52:10 crc kubenswrapper[4824]: I1209 09:52:10.265698 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"518105dd937b03952cbeccb9d4ed799027851c59f49636fa5a7e96cc5eeae762"} err="failed to get container status \"518105dd937b03952cbeccb9d4ed799027851c59f49636fa5a7e96cc5eeae762\": rpc error: code = NotFound desc = could not find container \"518105dd937b03952cbeccb9d4ed799027851c59f49636fa5a7e96cc5eeae762\": container with ID starting with 518105dd937b03952cbeccb9d4ed799027851c59f49636fa5a7e96cc5eeae762 not found: ID does not exist" Dec 09 09:52:10 crc kubenswrapper[4824]: I1209 09:52:10.359317 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 09 09:52:10 crc kubenswrapper[4824]: I1209 09:52:10.359403 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 09 09:52:10 crc kubenswrapper[4824]: I1209 09:52:10.359502 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 09 09:52:10 crc kubenswrapper[4824]: I1209 09:52:10.359570 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 09 09:52:10 crc kubenswrapper[4824]: I1209 09:52:10.359616 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 09 09:52:10 crc kubenswrapper[4824]: I1209 09:52:10.359631 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:52:10 crc kubenswrapper[4824]: I1209 09:52:10.359683 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:52:10 crc kubenswrapper[4824]: I1209 09:52:10.359766 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:52:10 crc kubenswrapper[4824]: I1209 09:52:10.359810 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:52:10 crc kubenswrapper[4824]: I1209 09:52:10.360403 4824 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 09 09:52:10 crc kubenswrapper[4824]: I1209 09:52:10.360429 4824 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 09 09:52:10 crc kubenswrapper[4824]: I1209 09:52:10.360448 4824 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 09 09:52:10 crc kubenswrapper[4824]: I1209 09:52:10.360461 4824 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 09 09:52:10 crc kubenswrapper[4824]: I1209 09:52:10.369806 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:52:10 crc kubenswrapper[4824]: I1209 09:52:10.462583 4824 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 09 09:52:11 crc kubenswrapper[4824]: I1209 09:52:11.917878 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 09 09:52:11 crc kubenswrapper[4824]: I1209 09:52:11.918225 4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 09 09:52:11 crc kubenswrapper[4824]: I1209 09:52:11.934428 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 09 09:52:11 crc kubenswrapper[4824]: I1209 09:52:11.934480 4824 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="a258de10-6378-479b-92da-a4c68ab5a5e2" Dec 09 09:52:11 crc kubenswrapper[4824]: I1209 09:52:11.940135 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 09 09:52:11 crc kubenswrapper[4824]: I1209 09:52:11.940219 4824 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="a258de10-6378-479b-92da-a4c68ab5a5e2" Dec 09 09:52:16 crc kubenswrapper[4824]: I1209 09:52:16.339064 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 09 09:52:21 crc kubenswrapper[4824]: I1209 09:52:21.046484 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 09 09:52:22 crc kubenswrapper[4824]: I1209 09:52:22.118095 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 09 09:52:23 crc kubenswrapper[4824]: I1209 09:52:23.798659 4824 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 09 09:52:24 crc kubenswrapper[4824]: I1209 09:52:24.810352 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 09 09:52:24 crc kubenswrapper[4824]: I1209 09:52:24.980734 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 09 09:52:25 crc kubenswrapper[4824]: I1209 09:52:25.267038 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 09 09:52:25 crc kubenswrapper[4824]: I1209 09:52:25.343591 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 09 09:52:25 crc kubenswrapper[4824]: I1209 09:52:25.347659 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 09 09:52:25 crc kubenswrapper[4824]: I1209 09:52:25.347727 4824 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="8b2a31f728d08d316ea20f26b82c72e791d8148fb0e65b5ccd6c881dfaf923e1" exitCode=137 Dec 09 09:52:25 crc kubenswrapper[4824]: I1209 09:52:25.347799 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"8b2a31f728d08d316ea20f26b82c72e791d8148fb0e65b5ccd6c881dfaf923e1"} Dec 09 09:52:25 crc kubenswrapper[4824]: I1209 09:52:25.347848 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"56e7587d374a363cda0fabf74a8b632232d4ba0c1278b72f249fc15e66a5e8eb"} Dec 09 09:52:25 crc kubenswrapper[4824]: I1209 09:52:25.347874 4824 scope.go:117] "RemoveContainer" containerID="98ddf8a4f94b8f9cec6f0ebadb61008f1afc04e5be1a4669e08bdb0bd3e2bdf0" Dec 09 09:52:25 crc kubenswrapper[4824]: I1209 09:52:25.619480 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 09 09:52:26 crc kubenswrapper[4824]: I1209 09:52:26.357108 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 09 09:52:27 crc kubenswrapper[4824]: I1209 09:52:27.000275 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 09 09:52:27 crc kubenswrapper[4824]: I1209 09:52:27.035063 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 09 09:52:27 crc kubenswrapper[4824]: I1209 09:52:27.131429 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 09 09:52:28 crc kubenswrapper[4824]: I1209 09:52:28.501747 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 09 09:52:29 crc kubenswrapper[4824]: I1209 09:52:29.463692 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 09 09:52:30 crc kubenswrapper[4824]: I1209 09:52:30.091575 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 09 09:52:32 crc kubenswrapper[4824]: I1209 09:52:32.008229 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 09 09:52:33 crc kubenswrapper[4824]: I1209 09:52:33.827424 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 09 09:52:34 crc kubenswrapper[4824]: I1209 09:52:34.312596 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 09 09:52:34 crc kubenswrapper[4824]: I1209 09:52:34.338345 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 09:52:34 crc kubenswrapper[4824]: I1209 09:52:34.936880 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 09:52:34 crc kubenswrapper[4824]: I1209 09:52:34.941834 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 09:52:35 crc kubenswrapper[4824]: I1209 09:52:35.105151 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 09 09:52:35 crc kubenswrapper[4824]: I1209 09:52:35.418602 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 09:52:35 crc kubenswrapper[4824]: I1209 09:52:35.451927 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 09 09:52:37 crc kubenswrapper[4824]: I1209 09:52:37.049376 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 09 09:52:37 crc kubenswrapper[4824]: I1209 09:52:37.373238 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 09 09:52:37 crc kubenswrapper[4824]: I1209 09:52:37.772856 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 09 09:52:39 crc kubenswrapper[4824]: I1209 09:52:39.412117 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 09 09:52:41 crc kubenswrapper[4824]: I1209 09:52:41.534153 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.289718 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-n8f5n"] Dec 09 09:52:42 crc kubenswrapper[4824]: E1209 09:52:42.290191 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.290219 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.290354 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.290844 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-n8f5n" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.293511 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-dockercfg-wwt9l" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.295400 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv"] Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.295765 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv" podUID="2d573cb8-34fa-40ac-98e6-c4f18caf4e0b" containerName="route-controller-manager" containerID="cri-o://9a26e71486ae07234e59f7f88717ebf731f51d54029951bf6107b3239eccb2a6" gracePeriod=30 Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.312084 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"telemetry-config" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.312460 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-root-ca.crt" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.312692 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"openshift-service-ca.crt" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.320250 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-tls" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.331739 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tqhnj"] Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.339674 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" podUID="b18b84da-1277-4a24-b434-c93eb83a3af8" containerName="controller-manager" containerID="cri-o://c5836d095b038fd556cbc9efa83d716da01139ebf660b51cb119df979bc7ef95" gracePeriod=30 Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.343856 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-n8f5n"] Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.348514 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xct2\" (UniqueName: \"kubernetes.io/projected/ff392a1f-7c4d-4f84-86ba-6ae01bcb9dea-kube-api-access-7xct2\") pod \"cluster-monitoring-operator-6d5b84845-n8f5n\" (UID: \"ff392a1f-7c4d-4f84-86ba-6ae01bcb9dea\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-n8f5n" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.348584 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/ff392a1f-7c4d-4f84-86ba-6ae01bcb9dea-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-n8f5n\" (UID: \"ff392a1f-7c4d-4f84-86ba-6ae01bcb9dea\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-n8f5n" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.348747 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/ff392a1f-7c4d-4f84-86ba-6ae01bcb9dea-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-n8f5n\" (UID: \"ff392a1f-7c4d-4f84-86ba-6ae01bcb9dea\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-n8f5n" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.449960 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xct2\" (UniqueName: \"kubernetes.io/projected/ff392a1f-7c4d-4f84-86ba-6ae01bcb9dea-kube-api-access-7xct2\") pod \"cluster-monitoring-operator-6d5b84845-n8f5n\" (UID: \"ff392a1f-7c4d-4f84-86ba-6ae01bcb9dea\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-n8f5n" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.450018 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/ff392a1f-7c4d-4f84-86ba-6ae01bcb9dea-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-n8f5n\" (UID: \"ff392a1f-7c4d-4f84-86ba-6ae01bcb9dea\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-n8f5n" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.450083 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/ff392a1f-7c4d-4f84-86ba-6ae01bcb9dea-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-n8f5n\" (UID: \"ff392a1f-7c4d-4f84-86ba-6ae01bcb9dea\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-n8f5n" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.452131 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/ff392a1f-7c4d-4f84-86ba-6ae01bcb9dea-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-n8f5n\" (UID: \"ff392a1f-7c4d-4f84-86ba-6ae01bcb9dea\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-n8f5n" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.471053 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/ff392a1f-7c4d-4f84-86ba-6ae01bcb9dea-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-n8f5n\" (UID: \"ff392a1f-7c4d-4f84-86ba-6ae01bcb9dea\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-n8f5n" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.499252 4824 generic.go:334] "Generic (PLEG): container finished" podID="2d573cb8-34fa-40ac-98e6-c4f18caf4e0b" containerID="9a26e71486ae07234e59f7f88717ebf731f51d54029951bf6107b3239eccb2a6" exitCode=0 Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.499844 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv" event={"ID":"2d573cb8-34fa-40ac-98e6-c4f18caf4e0b","Type":"ContainerDied","Data":"9a26e71486ae07234e59f7f88717ebf731f51d54029951bf6107b3239eccb2a6"} Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.502200 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xct2\" (UniqueName: \"kubernetes.io/projected/ff392a1f-7c4d-4f84-86ba-6ae01bcb9dea-kube-api-access-7xct2\") pod \"cluster-monitoring-operator-6d5b84845-n8f5n\" (UID: \"ff392a1f-7c4d-4f84-86ba-6ae01bcb9dea\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-n8f5n" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.619670 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-n8f5n" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.906046 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.942338 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.968562 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b18b84da-1277-4a24-b434-c93eb83a3af8-config\") pod \"b18b84da-1277-4a24-b434-c93eb83a3af8\" (UID: \"b18b84da-1277-4a24-b434-c93eb83a3af8\") " Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.968621 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b-serving-cert\") pod \"2d573cb8-34fa-40ac-98e6-c4f18caf4e0b\" (UID: \"2d573cb8-34fa-40ac-98e6-c4f18caf4e0b\") " Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.968677 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jh8zh\" (UniqueName: \"kubernetes.io/projected/b18b84da-1277-4a24-b434-c93eb83a3af8-kube-api-access-jh8zh\") pod \"b18b84da-1277-4a24-b434-c93eb83a3af8\" (UID: \"b18b84da-1277-4a24-b434-c93eb83a3af8\") " Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.968804 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b18b84da-1277-4a24-b434-c93eb83a3af8-serving-cert\") pod \"b18b84da-1277-4a24-b434-c93eb83a3af8\" (UID: \"b18b84da-1277-4a24-b434-c93eb83a3af8\") " Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.968824 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b18b84da-1277-4a24-b434-c93eb83a3af8-client-ca\") pod \"b18b84da-1277-4a24-b434-c93eb83a3af8\" (UID: \"b18b84da-1277-4a24-b434-c93eb83a3af8\") " Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.968881 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gh4vw\" (UniqueName: \"kubernetes.io/projected/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b-kube-api-access-gh4vw\") pod \"2d573cb8-34fa-40ac-98e6-c4f18caf4e0b\" (UID: \"2d573cb8-34fa-40ac-98e6-c4f18caf4e0b\") " Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.968910 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b-client-ca\") pod \"2d573cb8-34fa-40ac-98e6-c4f18caf4e0b\" (UID: \"2d573cb8-34fa-40ac-98e6-c4f18caf4e0b\") " Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.968937 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b-config\") pod \"2d573cb8-34fa-40ac-98e6-c4f18caf4e0b\" (UID: \"2d573cb8-34fa-40ac-98e6-c4f18caf4e0b\") " Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.968983 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b18b84da-1277-4a24-b434-c93eb83a3af8-proxy-ca-bundles\") pod \"b18b84da-1277-4a24-b434-c93eb83a3af8\" (UID: \"b18b84da-1277-4a24-b434-c93eb83a3af8\") " Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.969933 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b18b84da-1277-4a24-b434-c93eb83a3af8-config" (OuterVolumeSpecName: "config") pod "b18b84da-1277-4a24-b434-c93eb83a3af8" (UID: "b18b84da-1277-4a24-b434-c93eb83a3af8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.970436 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b-client-ca" (OuterVolumeSpecName: "client-ca") pod "2d573cb8-34fa-40ac-98e6-c4f18caf4e0b" (UID: "2d573cb8-34fa-40ac-98e6-c4f18caf4e0b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.971252 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b-config" (OuterVolumeSpecName: "config") pod "2d573cb8-34fa-40ac-98e6-c4f18caf4e0b" (UID: "2d573cb8-34fa-40ac-98e6-c4f18caf4e0b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.971870 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b18b84da-1277-4a24-b434-c93eb83a3af8-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b18b84da-1277-4a24-b434-c93eb83a3af8" (UID: "b18b84da-1277-4a24-b434-c93eb83a3af8"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.971938 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b18b84da-1277-4a24-b434-c93eb83a3af8-client-ca" (OuterVolumeSpecName: "client-ca") pod "b18b84da-1277-4a24-b434-c93eb83a3af8" (UID: "b18b84da-1277-4a24-b434-c93eb83a3af8"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.976268 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b-kube-api-access-gh4vw" (OuterVolumeSpecName: "kube-api-access-gh4vw") pod "2d573cb8-34fa-40ac-98e6-c4f18caf4e0b" (UID: "2d573cb8-34fa-40ac-98e6-c4f18caf4e0b"). InnerVolumeSpecName "kube-api-access-gh4vw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.976588 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b18b84da-1277-4a24-b434-c93eb83a3af8-kube-api-access-jh8zh" (OuterVolumeSpecName: "kube-api-access-jh8zh") pod "b18b84da-1277-4a24-b434-c93eb83a3af8" (UID: "b18b84da-1277-4a24-b434-c93eb83a3af8"). InnerVolumeSpecName "kube-api-access-jh8zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.978047 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2d573cb8-34fa-40ac-98e6-c4f18caf4e0b" (UID: "2d573cb8-34fa-40ac-98e6-c4f18caf4e0b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:52:42 crc kubenswrapper[4824]: I1209 09:52:42.981680 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b18b84da-1277-4a24-b434-c93eb83a3af8-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b18b84da-1277-4a24-b434-c93eb83a3af8" (UID: "b18b84da-1277-4a24-b434-c93eb83a3af8"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.070458 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.070970 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b18b84da-1277-4a24-b434-c93eb83a3af8-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.070987 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b18b84da-1277-4a24-b434-c93eb83a3af8-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.071002 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.071021 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jh8zh\" (UniqueName: \"kubernetes.io/projected/b18b84da-1277-4a24-b434-c93eb83a3af8-kube-api-access-jh8zh\") on node \"crc\" DevicePath \"\"" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.071036 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b18b84da-1277-4a24-b434-c93eb83a3af8-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.071053 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b18b84da-1277-4a24-b434-c93eb83a3af8-client-ca\") on node \"crc\" DevicePath \"\"" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.071068 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gh4vw\" (UniqueName: \"kubernetes.io/projected/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b-kube-api-access-gh4vw\") on node \"crc\" DevicePath \"\"" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.071087 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b-client-ca\") on node \"crc\" DevicePath \"\"" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.127551 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-n8f5n"] Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.506600 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-n8f5n" event={"ID":"ff392a1f-7c4d-4f84-86ba-6ae01bcb9dea","Type":"ContainerStarted","Data":"251a2b589dfe2b5dbf267f1496fae5d7e76f45e41b95801797ae2613d2a1b1e3"} Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.508398 4824 generic.go:334] "Generic (PLEG): container finished" podID="b18b84da-1277-4a24-b434-c93eb83a3af8" containerID="c5836d095b038fd556cbc9efa83d716da01139ebf660b51cb119df979bc7ef95" exitCode=0 Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.508455 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" event={"ID":"b18b84da-1277-4a24-b434-c93eb83a3af8","Type":"ContainerDied","Data":"c5836d095b038fd556cbc9efa83d716da01139ebf660b51cb119df979bc7ef95"} Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.508476 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" event={"ID":"b18b84da-1277-4a24-b434-c93eb83a3af8","Type":"ContainerDied","Data":"411a09bdb1a19b07848809653201eee2878a638679bbb1821a4de0a6cf2e9e89"} Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.508496 4824 scope.go:117] "RemoveContainer" containerID="c5836d095b038fd556cbc9efa83d716da01139ebf660b51cb119df979bc7ef95" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.508630 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tqhnj" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.520237 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv" event={"ID":"2d573cb8-34fa-40ac-98e6-c4f18caf4e0b","Type":"ContainerDied","Data":"dbccaeda174365dcbd986aed571d82d3572d104f23bc091f22eb0d456a26677c"} Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.520545 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.540716 4824 scope.go:117] "RemoveContainer" containerID="c5836d095b038fd556cbc9efa83d716da01139ebf660b51cb119df979bc7ef95" Dec 09 09:52:43 crc kubenswrapper[4824]: E1209 09:52:43.547961 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5836d095b038fd556cbc9efa83d716da01139ebf660b51cb119df979bc7ef95\": container with ID starting with c5836d095b038fd556cbc9efa83d716da01139ebf660b51cb119df979bc7ef95 not found: ID does not exist" containerID="c5836d095b038fd556cbc9efa83d716da01139ebf660b51cb119df979bc7ef95" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.548019 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5836d095b038fd556cbc9efa83d716da01139ebf660b51cb119df979bc7ef95"} err="failed to get container status \"c5836d095b038fd556cbc9efa83d716da01139ebf660b51cb119df979bc7ef95\": rpc error: code = NotFound desc = could not find container \"c5836d095b038fd556cbc9efa83d716da01139ebf660b51cb119df979bc7ef95\": container with ID starting with c5836d095b038fd556cbc9efa83d716da01139ebf660b51cb119df979bc7ef95 not found: ID does not exist" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.548058 4824 scope.go:117] "RemoveContainer" containerID="9a26e71486ae07234e59f7f88717ebf731f51d54029951bf6107b3239eccb2a6" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.548869 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tqhnj"] Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.556246 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tqhnj"] Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.570806 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv"] Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.575375 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqnfv"] Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.739687 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6ffd5d66b-c8452"] Dec 09 09:52:43 crc kubenswrapper[4824]: E1209 09:52:43.740691 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b18b84da-1277-4a24-b434-c93eb83a3af8" containerName="controller-manager" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.740713 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b18b84da-1277-4a24-b434-c93eb83a3af8" containerName="controller-manager" Dec 09 09:52:43 crc kubenswrapper[4824]: E1209 09:52:43.740741 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d573cb8-34fa-40ac-98e6-c4f18caf4e0b" containerName="route-controller-manager" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.740752 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d573cb8-34fa-40ac-98e6-c4f18caf4e0b" containerName="route-controller-manager" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.740949 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b18b84da-1277-4a24-b434-c93eb83a3af8" containerName="controller-manager" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.740972 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d573cb8-34fa-40ac-98e6-c4f18caf4e0b" containerName="route-controller-manager" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.741655 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.747010 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.748012 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.748065 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.748072 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.748156 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.748173 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.752227 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69"] Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.754222 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.756886 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.757164 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.757236 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.757264 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.757080 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.757109 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.761530 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.765051 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69"] Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.774772 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6ffd5d66b-c8452"] Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.782752 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bcaf42e7-e583-4767-ac14-a48f7160823a-serving-cert\") pod \"controller-manager-6ffd5d66b-c8452\" (UID: \"bcaf42e7-e583-4767-ac14-a48f7160823a\") " pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.782829 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd469\" (UniqueName: \"kubernetes.io/projected/bcaf42e7-e583-4767-ac14-a48f7160823a-kube-api-access-cd469\") pod \"controller-manager-6ffd5d66b-c8452\" (UID: \"bcaf42e7-e583-4767-ac14-a48f7160823a\") " pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.782908 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8dch\" (UniqueName: \"kubernetes.io/projected/47b1f313-33f1-4e82-8275-aeda6996d129-kube-api-access-z8dch\") pod \"route-controller-manager-57cffcc444-gxn69\" (UID: \"47b1f313-33f1-4e82-8275-aeda6996d129\") " pod="openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.782949 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47b1f313-33f1-4e82-8275-aeda6996d129-serving-cert\") pod \"route-controller-manager-57cffcc444-gxn69\" (UID: \"47b1f313-33f1-4e82-8275-aeda6996d129\") " pod="openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.783047 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcaf42e7-e583-4767-ac14-a48f7160823a-config\") pod \"controller-manager-6ffd5d66b-c8452\" (UID: \"bcaf42e7-e583-4767-ac14-a48f7160823a\") " pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.783115 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bcaf42e7-e583-4767-ac14-a48f7160823a-client-ca\") pod \"controller-manager-6ffd5d66b-c8452\" (UID: \"bcaf42e7-e583-4767-ac14-a48f7160823a\") " pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.783138 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bcaf42e7-e583-4767-ac14-a48f7160823a-proxy-ca-bundles\") pod \"controller-manager-6ffd5d66b-c8452\" (UID: \"bcaf42e7-e583-4767-ac14-a48f7160823a\") " pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.783180 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47b1f313-33f1-4e82-8275-aeda6996d129-client-ca\") pod \"route-controller-manager-57cffcc444-gxn69\" (UID: \"47b1f313-33f1-4e82-8275-aeda6996d129\") " pod="openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.783204 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47b1f313-33f1-4e82-8275-aeda6996d129-config\") pod \"route-controller-manager-57cffcc444-gxn69\" (UID: \"47b1f313-33f1-4e82-8275-aeda6996d129\") " pod="openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.885258 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47b1f313-33f1-4e82-8275-aeda6996d129-client-ca\") pod \"route-controller-manager-57cffcc444-gxn69\" (UID: \"47b1f313-33f1-4e82-8275-aeda6996d129\") " pod="openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.885326 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47b1f313-33f1-4e82-8275-aeda6996d129-config\") pod \"route-controller-manager-57cffcc444-gxn69\" (UID: \"47b1f313-33f1-4e82-8275-aeda6996d129\") " pod="openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.885382 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bcaf42e7-e583-4767-ac14-a48f7160823a-serving-cert\") pod \"controller-manager-6ffd5d66b-c8452\" (UID: \"bcaf42e7-e583-4767-ac14-a48f7160823a\") " pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.885418 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd469\" (UniqueName: \"kubernetes.io/projected/bcaf42e7-e583-4767-ac14-a48f7160823a-kube-api-access-cd469\") pod \"controller-manager-6ffd5d66b-c8452\" (UID: \"bcaf42e7-e583-4767-ac14-a48f7160823a\") " pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.885452 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8dch\" (UniqueName: \"kubernetes.io/projected/47b1f313-33f1-4e82-8275-aeda6996d129-kube-api-access-z8dch\") pod \"route-controller-manager-57cffcc444-gxn69\" (UID: \"47b1f313-33f1-4e82-8275-aeda6996d129\") " pod="openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.885480 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47b1f313-33f1-4e82-8275-aeda6996d129-serving-cert\") pod \"route-controller-manager-57cffcc444-gxn69\" (UID: \"47b1f313-33f1-4e82-8275-aeda6996d129\") " pod="openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.885497 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcaf42e7-e583-4767-ac14-a48f7160823a-config\") pod \"controller-manager-6ffd5d66b-c8452\" (UID: \"bcaf42e7-e583-4767-ac14-a48f7160823a\") " pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.885518 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bcaf42e7-e583-4767-ac14-a48f7160823a-client-ca\") pod \"controller-manager-6ffd5d66b-c8452\" (UID: \"bcaf42e7-e583-4767-ac14-a48f7160823a\") " pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.885557 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bcaf42e7-e583-4767-ac14-a48f7160823a-proxy-ca-bundles\") pod \"controller-manager-6ffd5d66b-c8452\" (UID: \"bcaf42e7-e583-4767-ac14-a48f7160823a\") " pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.886829 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47b1f313-33f1-4e82-8275-aeda6996d129-client-ca\") pod \"route-controller-manager-57cffcc444-gxn69\" (UID: \"47b1f313-33f1-4e82-8275-aeda6996d129\") " pod="openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.887043 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47b1f313-33f1-4e82-8275-aeda6996d129-config\") pod \"route-controller-manager-57cffcc444-gxn69\" (UID: \"47b1f313-33f1-4e82-8275-aeda6996d129\") " pod="openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.887287 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bcaf42e7-e583-4767-ac14-a48f7160823a-proxy-ca-bundles\") pod \"controller-manager-6ffd5d66b-c8452\" (UID: \"bcaf42e7-e583-4767-ac14-a48f7160823a\") " pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.887323 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bcaf42e7-e583-4767-ac14-a48f7160823a-client-ca\") pod \"controller-manager-6ffd5d66b-c8452\" (UID: \"bcaf42e7-e583-4767-ac14-a48f7160823a\") " pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.888124 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcaf42e7-e583-4767-ac14-a48f7160823a-config\") pod \"controller-manager-6ffd5d66b-c8452\" (UID: \"bcaf42e7-e583-4767-ac14-a48f7160823a\") " pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.895133 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47b1f313-33f1-4e82-8275-aeda6996d129-serving-cert\") pod \"route-controller-manager-57cffcc444-gxn69\" (UID: \"47b1f313-33f1-4e82-8275-aeda6996d129\") " pod="openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.896052 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bcaf42e7-e583-4767-ac14-a48f7160823a-serving-cert\") pod \"controller-manager-6ffd5d66b-c8452\" (UID: \"bcaf42e7-e583-4767-ac14-a48f7160823a\") " pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.902386 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd469\" (UniqueName: \"kubernetes.io/projected/bcaf42e7-e583-4767-ac14-a48f7160823a-kube-api-access-cd469\") pod \"controller-manager-6ffd5d66b-c8452\" (UID: \"bcaf42e7-e583-4767-ac14-a48f7160823a\") " pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.905410 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8dch\" (UniqueName: \"kubernetes.io/projected/47b1f313-33f1-4e82-8275-aeda6996d129-kube-api-access-z8dch\") pod \"route-controller-manager-57cffcc444-gxn69\" (UID: \"47b1f313-33f1-4e82-8275-aeda6996d129\") " pod="openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.925318 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d573cb8-34fa-40ac-98e6-c4f18caf4e0b" path="/var/lib/kubelet/pods/2d573cb8-34fa-40ac-98e6-c4f18caf4e0b/volumes" Dec 09 09:52:43 crc kubenswrapper[4824]: I1209 09:52:43.927257 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b18b84da-1277-4a24-b434-c93eb83a3af8" path="/var/lib/kubelet/pods/b18b84da-1277-4a24-b434-c93eb83a3af8/volumes" Dec 09 09:52:44 crc kubenswrapper[4824]: I1209 09:52:44.070709 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" Dec 09 09:52:44 crc kubenswrapper[4824]: I1209 09:52:44.084502 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69" Dec 09 09:52:44 crc kubenswrapper[4824]: I1209 09:52:44.565089 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6ffd5d66b-c8452"] Dec 09 09:52:44 crc kubenswrapper[4824]: I1209 09:52:44.570400 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69"] Dec 09 09:52:45 crc kubenswrapper[4824]: I1209 09:52:45.545291 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" event={"ID":"bcaf42e7-e583-4767-ac14-a48f7160823a","Type":"ContainerStarted","Data":"d237271ae948f8b60cb86d6c32f4ae74e4e577c581f4511931de72d6973f936e"} Dec 09 09:52:45 crc kubenswrapper[4824]: I1209 09:52:45.546116 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" event={"ID":"bcaf42e7-e583-4767-ac14-a48f7160823a","Type":"ContainerStarted","Data":"0db146f95ea4b3abc52c1489781f10403aa73f710724aa5d09ca689f7f081156"} Dec 09 09:52:45 crc kubenswrapper[4824]: I1209 09:52:45.547936 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" Dec 09 09:52:45 crc kubenswrapper[4824]: I1209 09:52:45.554749 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" Dec 09 09:52:45 crc kubenswrapper[4824]: I1209 09:52:45.558815 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69" event={"ID":"47b1f313-33f1-4e82-8275-aeda6996d129","Type":"ContainerStarted","Data":"33923f4b5efac3a4263403c1903ce8bc3daf06e00dddcbbb22991de702b5f803"} Dec 09 09:52:45 crc kubenswrapper[4824]: I1209 09:52:45.558872 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69" event={"ID":"47b1f313-33f1-4e82-8275-aeda6996d129","Type":"ContainerStarted","Data":"ead6f530a4fe812a1d6bbb1bca39d30a240c4bb297d7803df12f3c1f131ba8c6"} Dec 09 09:52:45 crc kubenswrapper[4824]: I1209 09:52:45.559341 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69" Dec 09 09:52:45 crc kubenswrapper[4824]: I1209 09:52:45.566804 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69" Dec 09 09:52:45 crc kubenswrapper[4824]: I1209 09:52:45.578942 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" podStartSLOduration=3.578903413 podStartE2EDuration="3.578903413s" podCreationTimestamp="2025-12-09 09:52:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:52:45.577615162 +0000 UTC m=+321.912119849" watchObservedRunningTime="2025-12-09 09:52:45.578903413 +0000 UTC m=+321.913408080" Dec 09 09:52:45 crc kubenswrapper[4824]: I1209 09:52:45.595853 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69"] Dec 09 09:52:45 crc kubenswrapper[4824]: I1209 09:52:45.630166 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69" podStartSLOduration=3.630137673 podStartE2EDuration="3.630137673s" podCreationTimestamp="2025-12-09 09:52:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:52:45.625408084 +0000 UTC m=+321.959912751" watchObservedRunningTime="2025-12-09 09:52:45.630137673 +0000 UTC m=+321.964642340" Dec 09 09:52:46 crc kubenswrapper[4824]: I1209 09:52:46.567233 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-n8f5n" event={"ID":"ff392a1f-7c4d-4f84-86ba-6ae01bcb9dea","Type":"ContainerStarted","Data":"d39822161061e7a7e3347da0d65c34e6258adb7563c8aa30fc79daea5714344c"} Dec 09 09:52:46 crc kubenswrapper[4824]: I1209 09:52:46.590990 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-n8f5n" podStartSLOduration=1.512775408 podStartE2EDuration="4.590962729s" podCreationTimestamp="2025-12-09 09:52:42 +0000 UTC" firstStartedPulling="2025-12-09 09:52:43.138372018 +0000 UTC m=+319.472876685" lastFinishedPulling="2025-12-09 09:52:46.216559339 +0000 UTC m=+322.551064006" observedRunningTime="2025-12-09 09:52:46.586247341 +0000 UTC m=+322.920752028" watchObservedRunningTime="2025-12-09 09:52:46.590962729 +0000 UTC m=+322.925467396" Dec 09 09:52:46 crc kubenswrapper[4824]: I1209 09:52:46.980447 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng"] Dec 09 09:52:46 crc kubenswrapper[4824]: I1209 09:52:46.981632 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" Dec 09 09:52:46 crc kubenswrapper[4824]: I1209 09:52:46.984379 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-tls" Dec 09 09:52:46 crc kubenswrapper[4824]: I1209 09:52:46.985881 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-dockercfg-btcdk" Dec 09 09:52:46 crc kubenswrapper[4824]: I1209 09:52:46.995467 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng"] Dec 09 09:52:47 crc kubenswrapper[4824]: I1209 09:52:47.077155 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/86c42164-acb8-4a27-9a0f-48307cd304e4-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-4dlng\" (UID: \"86c42164-acb8-4a27-9a0f-48307cd304e4\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" Dec 09 09:52:47 crc kubenswrapper[4824]: I1209 09:52:47.178640 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/86c42164-acb8-4a27-9a0f-48307cd304e4-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-4dlng\" (UID: \"86c42164-acb8-4a27-9a0f-48307cd304e4\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" Dec 09 09:52:47 crc kubenswrapper[4824]: I1209 09:52:47.196216 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/86c42164-acb8-4a27-9a0f-48307cd304e4-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-4dlng\" (UID: \"86c42164-acb8-4a27-9a0f-48307cd304e4\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" Dec 09 09:52:47 crc kubenswrapper[4824]: I1209 09:52:47.297904 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" Dec 09 09:52:47 crc kubenswrapper[4824]: I1209 09:52:47.574683 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69" podUID="47b1f313-33f1-4e82-8275-aeda6996d129" containerName="route-controller-manager" containerID="cri-o://33923f4b5efac3a4263403c1903ce8bc3daf06e00dddcbbb22991de702b5f803" gracePeriod=30 Dec 09 09:52:47 crc kubenswrapper[4824]: I1209 09:52:47.808258 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng"] Dec 09 09:52:47 crc kubenswrapper[4824]: I1209 09:52:47.958078 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69" Dec 09 09:52:47 crc kubenswrapper[4824]: I1209 09:52:47.986918 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9"] Dec 09 09:52:47 crc kubenswrapper[4824]: E1209 09:52:47.987228 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47b1f313-33f1-4e82-8275-aeda6996d129" containerName="route-controller-manager" Dec 09 09:52:47 crc kubenswrapper[4824]: I1209 09:52:47.987241 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="47b1f313-33f1-4e82-8275-aeda6996d129" containerName="route-controller-manager" Dec 09 09:52:47 crc kubenswrapper[4824]: I1209 09:52:47.987366 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="47b1f313-33f1-4e82-8275-aeda6996d129" containerName="route-controller-manager" Dec 09 09:52:47 crc kubenswrapper[4824]: I1209 09:52:47.987893 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.001767 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9"] Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.094897 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47b1f313-33f1-4e82-8275-aeda6996d129-serving-cert\") pod \"47b1f313-33f1-4e82-8275-aeda6996d129\" (UID: \"47b1f313-33f1-4e82-8275-aeda6996d129\") " Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.094986 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47b1f313-33f1-4e82-8275-aeda6996d129-client-ca\") pod \"47b1f313-33f1-4e82-8275-aeda6996d129\" (UID: \"47b1f313-33f1-4e82-8275-aeda6996d129\") " Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.095292 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47b1f313-33f1-4e82-8275-aeda6996d129-config\") pod \"47b1f313-33f1-4e82-8275-aeda6996d129\" (UID: \"47b1f313-33f1-4e82-8275-aeda6996d129\") " Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.095576 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8dch\" (UniqueName: \"kubernetes.io/projected/47b1f313-33f1-4e82-8275-aeda6996d129-kube-api-access-z8dch\") pod \"47b1f313-33f1-4e82-8275-aeda6996d129\" (UID: \"47b1f313-33f1-4e82-8275-aeda6996d129\") " Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.096067 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afe92d2b-bce3-445e-9c26-9c533353e7a3-config\") pod \"route-controller-manager-7f84c5b8f9-rqmr9\" (UID: \"afe92d2b-bce3-445e-9c26-9c533353e7a3\") " pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.096072 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47b1f313-33f1-4e82-8275-aeda6996d129-client-ca" (OuterVolumeSpecName: "client-ca") pod "47b1f313-33f1-4e82-8275-aeda6996d129" (UID: "47b1f313-33f1-4e82-8275-aeda6996d129"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.096179 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47b1f313-33f1-4e82-8275-aeda6996d129-config" (OuterVolumeSpecName: "config") pod "47b1f313-33f1-4e82-8275-aeda6996d129" (UID: "47b1f313-33f1-4e82-8275-aeda6996d129"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.096244 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/afe92d2b-bce3-445e-9c26-9c533353e7a3-client-ca\") pod \"route-controller-manager-7f84c5b8f9-rqmr9\" (UID: \"afe92d2b-bce3-445e-9c26-9c533353e7a3\") " pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.096368 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlmfw\" (UniqueName: \"kubernetes.io/projected/afe92d2b-bce3-445e-9c26-9c533353e7a3-kube-api-access-hlmfw\") pod \"route-controller-manager-7f84c5b8f9-rqmr9\" (UID: \"afe92d2b-bce3-445e-9c26-9c533353e7a3\") " pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.096522 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/afe92d2b-bce3-445e-9c26-9c533353e7a3-serving-cert\") pod \"route-controller-manager-7f84c5b8f9-rqmr9\" (UID: \"afe92d2b-bce3-445e-9c26-9c533353e7a3\") " pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.096709 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47b1f313-33f1-4e82-8275-aeda6996d129-client-ca\") on node \"crc\" DevicePath \"\"" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.096731 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47b1f313-33f1-4e82-8275-aeda6996d129-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.101762 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47b1f313-33f1-4e82-8275-aeda6996d129-kube-api-access-z8dch" (OuterVolumeSpecName: "kube-api-access-z8dch") pod "47b1f313-33f1-4e82-8275-aeda6996d129" (UID: "47b1f313-33f1-4e82-8275-aeda6996d129"). InnerVolumeSpecName "kube-api-access-z8dch". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.102501 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47b1f313-33f1-4e82-8275-aeda6996d129-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "47b1f313-33f1-4e82-8275-aeda6996d129" (UID: "47b1f313-33f1-4e82-8275-aeda6996d129"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.199196 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/afe92d2b-bce3-445e-9c26-9c533353e7a3-serving-cert\") pod \"route-controller-manager-7f84c5b8f9-rqmr9\" (UID: \"afe92d2b-bce3-445e-9c26-9c533353e7a3\") " pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.199332 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afe92d2b-bce3-445e-9c26-9c533353e7a3-config\") pod \"route-controller-manager-7f84c5b8f9-rqmr9\" (UID: \"afe92d2b-bce3-445e-9c26-9c533353e7a3\") " pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.199384 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/afe92d2b-bce3-445e-9c26-9c533353e7a3-client-ca\") pod \"route-controller-manager-7f84c5b8f9-rqmr9\" (UID: \"afe92d2b-bce3-445e-9c26-9c533353e7a3\") " pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.199442 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlmfw\" (UniqueName: \"kubernetes.io/projected/afe92d2b-bce3-445e-9c26-9c533353e7a3-kube-api-access-hlmfw\") pod \"route-controller-manager-7f84c5b8f9-rqmr9\" (UID: \"afe92d2b-bce3-445e-9c26-9c533353e7a3\") " pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.199518 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47b1f313-33f1-4e82-8275-aeda6996d129-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.199539 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8dch\" (UniqueName: \"kubernetes.io/projected/47b1f313-33f1-4e82-8275-aeda6996d129-kube-api-access-z8dch\") on node \"crc\" DevicePath \"\"" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.201262 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/afe92d2b-bce3-445e-9c26-9c533353e7a3-client-ca\") pod \"route-controller-manager-7f84c5b8f9-rqmr9\" (UID: \"afe92d2b-bce3-445e-9c26-9c533353e7a3\") " pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.201402 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afe92d2b-bce3-445e-9c26-9c533353e7a3-config\") pod \"route-controller-manager-7f84c5b8f9-rqmr9\" (UID: \"afe92d2b-bce3-445e-9c26-9c533353e7a3\") " pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.205080 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/afe92d2b-bce3-445e-9c26-9c533353e7a3-serving-cert\") pod \"route-controller-manager-7f84c5b8f9-rqmr9\" (UID: \"afe92d2b-bce3-445e-9c26-9c533353e7a3\") " pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.218470 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlmfw\" (UniqueName: \"kubernetes.io/projected/afe92d2b-bce3-445e-9c26-9c533353e7a3-kube-api-access-hlmfw\") pod \"route-controller-manager-7f84c5b8f9-rqmr9\" (UID: \"afe92d2b-bce3-445e-9c26-9c533353e7a3\") " pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.303869 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.583635 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" event={"ID":"86c42164-acb8-4a27-9a0f-48307cd304e4","Type":"ContainerStarted","Data":"f188766d4cafc88bc2f39c09465a0331212d26400fa8c0f26633d3e590202e68"} Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.585998 4824 generic.go:334] "Generic (PLEG): container finished" podID="47b1f313-33f1-4e82-8275-aeda6996d129" containerID="33923f4b5efac3a4263403c1903ce8bc3daf06e00dddcbbb22991de702b5f803" exitCode=0 Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.586043 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69" event={"ID":"47b1f313-33f1-4e82-8275-aeda6996d129","Type":"ContainerDied","Data":"33923f4b5efac3a4263403c1903ce8bc3daf06e00dddcbbb22991de702b5f803"} Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.586094 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69" event={"ID":"47b1f313-33f1-4e82-8275-aeda6996d129","Type":"ContainerDied","Data":"ead6f530a4fe812a1d6bbb1bca39d30a240c4bb297d7803df12f3c1f131ba8c6"} Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.586095 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.586121 4824 scope.go:117] "RemoveContainer" containerID="33923f4b5efac3a4263403c1903ce8bc3daf06e00dddcbbb22991de702b5f803" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.610487 4824 scope.go:117] "RemoveContainer" containerID="33923f4b5efac3a4263403c1903ce8bc3daf06e00dddcbbb22991de702b5f803" Dec 09 09:52:48 crc kubenswrapper[4824]: E1209 09:52:48.611316 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33923f4b5efac3a4263403c1903ce8bc3daf06e00dddcbbb22991de702b5f803\": container with ID starting with 33923f4b5efac3a4263403c1903ce8bc3daf06e00dddcbbb22991de702b5f803 not found: ID does not exist" containerID="33923f4b5efac3a4263403c1903ce8bc3daf06e00dddcbbb22991de702b5f803" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.611377 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33923f4b5efac3a4263403c1903ce8bc3daf06e00dddcbbb22991de702b5f803"} err="failed to get container status \"33923f4b5efac3a4263403c1903ce8bc3daf06e00dddcbbb22991de702b5f803\": rpc error: code = NotFound desc = could not find container \"33923f4b5efac3a4263403c1903ce8bc3daf06e00dddcbbb22991de702b5f803\": container with ID starting with 33923f4b5efac3a4263403c1903ce8bc3daf06e00dddcbbb22991de702b5f803 not found: ID does not exist" Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.624365 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69"] Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.627437 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-57cffcc444-gxn69"] Dec 09 09:52:48 crc kubenswrapper[4824]: I1209 09:52:48.733594 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9"] Dec 09 09:52:48 crc kubenswrapper[4824]: W1209 09:52:48.741419 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podafe92d2b_bce3_445e_9c26_9c533353e7a3.slice/crio-763f62fcd7468c578d0b8593abb21b54acd0e3beeebf600aa88c03111714b519 WatchSource:0}: Error finding container 763f62fcd7468c578d0b8593abb21b54acd0e3beeebf600aa88c03111714b519: Status 404 returned error can't find the container with id 763f62fcd7468c578d0b8593abb21b54acd0e3beeebf600aa88c03111714b519 Dec 09 09:52:49 crc kubenswrapper[4824]: I1209 09:52:49.597678 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" event={"ID":"afe92d2b-bce3-445e-9c26-9c533353e7a3","Type":"ContainerStarted","Data":"03c338c0a6a98a41c144c02fffabcfe6e5a3daf642e4e2523e58122af8753be4"} Dec 09 09:52:49 crc kubenswrapper[4824]: I1209 09:52:49.598166 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" event={"ID":"afe92d2b-bce3-445e-9c26-9c533353e7a3","Type":"ContainerStarted","Data":"763f62fcd7468c578d0b8593abb21b54acd0e3beeebf600aa88c03111714b519"} Dec 09 09:52:49 crc kubenswrapper[4824]: I1209 09:52:49.598193 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" Dec 09 09:52:49 crc kubenswrapper[4824]: I1209 09:52:49.605104 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" Dec 09 09:52:49 crc kubenswrapper[4824]: I1209 09:52:49.632694 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" podStartSLOduration=4.632667062 podStartE2EDuration="4.632667062s" podCreationTimestamp="2025-12-09 09:52:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:52:49.62942864 +0000 UTC m=+325.963933307" watchObservedRunningTime="2025-12-09 09:52:49.632667062 +0000 UTC m=+325.967171729" Dec 09 09:52:49 crc kubenswrapper[4824]: I1209 09:52:49.918056 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47b1f313-33f1-4e82-8275-aeda6996d129" path="/var/lib/kubelet/pods/47b1f313-33f1-4e82-8275-aeda6996d129/volumes" Dec 09 09:52:50 crc kubenswrapper[4824]: I1209 09:52:50.605726 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" event={"ID":"86c42164-acb8-4a27-9a0f-48307cd304e4","Type":"ContainerStarted","Data":"00bae8ec6f6e4cf5dd4670db60b6420c115bdac303b2c253cc78e718e60f743c"} Dec 09 09:52:50 crc kubenswrapper[4824]: I1209 09:52:50.606164 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" Dec 09 09:52:50 crc kubenswrapper[4824]: I1209 09:52:50.612398 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" Dec 09 09:52:50 crc kubenswrapper[4824]: I1209 09:52:50.630799 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" podStartSLOduration=2.737802159 podStartE2EDuration="4.630763079s" podCreationTimestamp="2025-12-09 09:52:46 +0000 UTC" firstStartedPulling="2025-12-09 09:52:47.845080795 +0000 UTC m=+324.179585462" lastFinishedPulling="2025-12-09 09:52:49.738041715 +0000 UTC m=+326.072546382" observedRunningTime="2025-12-09 09:52:50.629385547 +0000 UTC m=+326.963890214" watchObservedRunningTime="2025-12-09 09:52:50.630763079 +0000 UTC m=+326.965267746" Dec 09 09:52:51 crc kubenswrapper[4824]: I1209 09:52:51.061850 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-l2xq2"] Dec 09 09:52:51 crc kubenswrapper[4824]: I1209 09:52:51.063216 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-db54df47d-l2xq2" Dec 09 09:52:51 crc kubenswrapper[4824]: I1209 09:52:51.066589 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-kube-rbac-proxy-config" Dec 09 09:52:51 crc kubenswrapper[4824]: I1209 09:52:51.067336 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-tls" Dec 09 09:52:51 crc kubenswrapper[4824]: I1209 09:52:51.067689 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-dockercfg-7ckvr" Dec 09 09:52:51 crc kubenswrapper[4824]: I1209 09:52:51.067873 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-client-ca" Dec 09 09:52:51 crc kubenswrapper[4824]: I1209 09:52:51.085476 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-l2xq2"] Dec 09 09:52:51 crc kubenswrapper[4824]: I1209 09:52:51.194209 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/833a782e-9385-4695-ba79-dc23fbb8dacb-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-l2xq2\" (UID: \"833a782e-9385-4695-ba79-dc23fbb8dacb\") " pod="openshift-monitoring/prometheus-operator-db54df47d-l2xq2" Dec 09 09:52:51 crc kubenswrapper[4824]: I1209 09:52:51.194676 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/833a782e-9385-4695-ba79-dc23fbb8dacb-metrics-client-ca\") pod \"prometheus-operator-db54df47d-l2xq2\" (UID: \"833a782e-9385-4695-ba79-dc23fbb8dacb\") " pod="openshift-monitoring/prometheus-operator-db54df47d-l2xq2" Dec 09 09:52:51 crc kubenswrapper[4824]: I1209 09:52:51.194846 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/833a782e-9385-4695-ba79-dc23fbb8dacb-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-l2xq2\" (UID: \"833a782e-9385-4695-ba79-dc23fbb8dacb\") " pod="openshift-monitoring/prometheus-operator-db54df47d-l2xq2" Dec 09 09:52:51 crc kubenswrapper[4824]: I1209 09:52:51.194895 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqx9l\" (UniqueName: \"kubernetes.io/projected/833a782e-9385-4695-ba79-dc23fbb8dacb-kube-api-access-hqx9l\") pod \"prometheus-operator-db54df47d-l2xq2\" (UID: \"833a782e-9385-4695-ba79-dc23fbb8dacb\") " pod="openshift-monitoring/prometheus-operator-db54df47d-l2xq2" Dec 09 09:52:51 crc kubenswrapper[4824]: I1209 09:52:51.296203 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/833a782e-9385-4695-ba79-dc23fbb8dacb-metrics-client-ca\") pod \"prometheus-operator-db54df47d-l2xq2\" (UID: \"833a782e-9385-4695-ba79-dc23fbb8dacb\") " pod="openshift-monitoring/prometheus-operator-db54df47d-l2xq2" Dec 09 09:52:51 crc kubenswrapper[4824]: I1209 09:52:51.296289 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/833a782e-9385-4695-ba79-dc23fbb8dacb-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-l2xq2\" (UID: \"833a782e-9385-4695-ba79-dc23fbb8dacb\") " pod="openshift-monitoring/prometheus-operator-db54df47d-l2xq2" Dec 09 09:52:51 crc kubenswrapper[4824]: I1209 09:52:51.296311 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqx9l\" (UniqueName: \"kubernetes.io/projected/833a782e-9385-4695-ba79-dc23fbb8dacb-kube-api-access-hqx9l\") pod \"prometheus-operator-db54df47d-l2xq2\" (UID: \"833a782e-9385-4695-ba79-dc23fbb8dacb\") " pod="openshift-monitoring/prometheus-operator-db54df47d-l2xq2" Dec 09 09:52:51 crc kubenswrapper[4824]: I1209 09:52:51.296351 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/833a782e-9385-4695-ba79-dc23fbb8dacb-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-l2xq2\" (UID: \"833a782e-9385-4695-ba79-dc23fbb8dacb\") " pod="openshift-monitoring/prometheus-operator-db54df47d-l2xq2" Dec 09 09:52:51 crc kubenswrapper[4824]: I1209 09:52:51.297341 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/833a782e-9385-4695-ba79-dc23fbb8dacb-metrics-client-ca\") pod \"prometheus-operator-db54df47d-l2xq2\" (UID: \"833a782e-9385-4695-ba79-dc23fbb8dacb\") " pod="openshift-monitoring/prometheus-operator-db54df47d-l2xq2" Dec 09 09:52:51 crc kubenswrapper[4824]: I1209 09:52:51.303924 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/833a782e-9385-4695-ba79-dc23fbb8dacb-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-l2xq2\" (UID: \"833a782e-9385-4695-ba79-dc23fbb8dacb\") " pod="openshift-monitoring/prometheus-operator-db54df47d-l2xq2" Dec 09 09:52:51 crc kubenswrapper[4824]: I1209 09:52:51.315146 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqx9l\" (UniqueName: \"kubernetes.io/projected/833a782e-9385-4695-ba79-dc23fbb8dacb-kube-api-access-hqx9l\") pod \"prometheus-operator-db54df47d-l2xq2\" (UID: \"833a782e-9385-4695-ba79-dc23fbb8dacb\") " pod="openshift-monitoring/prometheus-operator-db54df47d-l2xq2" Dec 09 09:52:51 crc kubenswrapper[4824]: I1209 09:52:51.318345 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/833a782e-9385-4695-ba79-dc23fbb8dacb-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-l2xq2\" (UID: \"833a782e-9385-4695-ba79-dc23fbb8dacb\") " pod="openshift-monitoring/prometheus-operator-db54df47d-l2xq2" Dec 09 09:52:51 crc kubenswrapper[4824]: I1209 09:52:51.384113 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-db54df47d-l2xq2" Dec 09 09:52:51 crc kubenswrapper[4824]: I1209 09:52:51.921035 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-l2xq2"] Dec 09 09:52:52 crc kubenswrapper[4824]: I1209 09:52:52.621509 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-l2xq2" event={"ID":"833a782e-9385-4695-ba79-dc23fbb8dacb","Type":"ContainerStarted","Data":"8adef94ae5e35211238c9592303e4da02ef01240bdfe298183b582a3253ee82f"} Dec 09 09:52:54 crc kubenswrapper[4824]: I1209 09:52:54.637889 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-l2xq2" event={"ID":"833a782e-9385-4695-ba79-dc23fbb8dacb","Type":"ContainerStarted","Data":"41b780a202b249844157d269140af2017e3cfacd98c49ce56d5541e1c7b17497"} Dec 09 09:52:54 crc kubenswrapper[4824]: I1209 09:52:54.638358 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-l2xq2" event={"ID":"833a782e-9385-4695-ba79-dc23fbb8dacb","Type":"ContainerStarted","Data":"59b2808b5a6b6008a590da8fc585f512c3a4e556228ee82e036a3b6116759a64"} Dec 09 09:52:54 crc kubenswrapper[4824]: I1209 09:52:54.660466 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-db54df47d-l2xq2" podStartSLOduration=1.339755819 podStartE2EDuration="3.6604376s" podCreationTimestamp="2025-12-09 09:52:51 +0000 UTC" firstStartedPulling="2025-12-09 09:52:51.921737724 +0000 UTC m=+328.256242381" lastFinishedPulling="2025-12-09 09:52:54.242419495 +0000 UTC m=+330.576924162" observedRunningTime="2025-12-09 09:52:54.657279771 +0000 UTC m=+330.991784438" watchObservedRunningTime="2025-12-09 09:52:54.6604376 +0000 UTC m=+330.994942267" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.479418 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-hjnrk"] Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.481456 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-566fddb674-hjnrk" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.486093 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-kube-rbac-proxy-config" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.486163 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-tls" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.486414 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-dockercfg-nbtvk" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.504715 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-hjnrk"] Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.549413 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw"] Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.550662 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-78nq2"] Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.554702 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.555356 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.563741 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-state-metrics-custom-resource-state-configmap" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.564043 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-dockercfg-df9dt" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.564196 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-kube-rbac-proxy-config" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.565039 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-tls" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.565160 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-tls" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.565191 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw"] Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.565277 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-kube-rbac-proxy-config" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.565390 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-dockercfg-q4tn9" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.580474 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/f2bcdf28-cd59-4f1f-b55d-614395ff2788-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-hjnrk\" (UID: \"f2bcdf28-cd59-4f1f-b55d-614395ff2788\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hjnrk" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.580562 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/f2bcdf28-cd59-4f1f-b55d-614395ff2788-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-hjnrk\" (UID: \"f2bcdf28-cd59-4f1f-b55d-614395ff2788\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hjnrk" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.580616 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8w2d\" (UniqueName: \"kubernetes.io/projected/f2bcdf28-cd59-4f1f-b55d-614395ff2788-kube-api-access-t8w2d\") pod \"openshift-state-metrics-566fddb674-hjnrk\" (UID: \"f2bcdf28-cd59-4f1f-b55d-614395ff2788\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hjnrk" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.580670 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/f2bcdf28-cd59-4f1f-b55d-614395ff2788-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-hjnrk\" (UID: \"f2bcdf28-cd59-4f1f-b55d-614395ff2788\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hjnrk" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.682555 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c650b71d-b6e9-4323-b8e0-52e335852881-sys\") pod \"node-exporter-78nq2\" (UID: \"c650b71d-b6e9-4323-b8e0-52e335852881\") " pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.682646 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/c650b71d-b6e9-4323-b8e0-52e335852881-node-exporter-textfile\") pod \"node-exporter-78nq2\" (UID: \"c650b71d-b6e9-4323-b8e0-52e335852881\") " pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.682692 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/c650b71d-b6e9-4323-b8e0-52e335852881-root\") pod \"node-exporter-78nq2\" (UID: \"c650b71d-b6e9-4323-b8e0-52e335852881\") " pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.682718 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/c650b71d-b6e9-4323-b8e0-52e335852881-node-exporter-wtmp\") pod \"node-exporter-78nq2\" (UID: \"c650b71d-b6e9-4323-b8e0-52e335852881\") " pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.682765 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/f2bcdf28-cd59-4f1f-b55d-614395ff2788-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-hjnrk\" (UID: \"f2bcdf28-cd59-4f1f-b55d-614395ff2788\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hjnrk" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.682822 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/f2bcdf28-cd59-4f1f-b55d-614395ff2788-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-hjnrk\" (UID: \"f2bcdf28-cd59-4f1f-b55d-614395ff2788\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hjnrk" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.682851 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/aea84e78-a09c-400c-9e99-d5e7aca0ce51-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-tdzdw\" (UID: \"aea84e78-a09c-400c-9e99-d5e7aca0ce51\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.682883 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/c650b71d-b6e9-4323-b8e0-52e335852881-metrics-client-ca\") pod \"node-exporter-78nq2\" (UID: \"c650b71d-b6e9-4323-b8e0-52e335852881\") " pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.682919 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8w2d\" (UniqueName: \"kubernetes.io/projected/f2bcdf28-cd59-4f1f-b55d-614395ff2788-kube-api-access-t8w2d\") pod \"openshift-state-metrics-566fddb674-hjnrk\" (UID: \"f2bcdf28-cd59-4f1f-b55d-614395ff2788\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hjnrk" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.682959 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/c650b71d-b6e9-4323-b8e0-52e335852881-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-78nq2\" (UID: \"c650b71d-b6e9-4323-b8e0-52e335852881\") " pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.682993 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/aea84e78-a09c-400c-9e99-d5e7aca0ce51-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-tdzdw\" (UID: \"aea84e78-a09c-400c-9e99-d5e7aca0ce51\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.683023 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcd94\" (UniqueName: \"kubernetes.io/projected/aea84e78-a09c-400c-9e99-d5e7aca0ce51-kube-api-access-lcd94\") pod \"kube-state-metrics-777cb5bd5d-tdzdw\" (UID: \"aea84e78-a09c-400c-9e99-d5e7aca0ce51\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.683048 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77gln\" (UniqueName: \"kubernetes.io/projected/c650b71d-b6e9-4323-b8e0-52e335852881-kube-api-access-77gln\") pod \"node-exporter-78nq2\" (UID: \"c650b71d-b6e9-4323-b8e0-52e335852881\") " pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.683070 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/c650b71d-b6e9-4323-b8e0-52e335852881-node-exporter-tls\") pod \"node-exporter-78nq2\" (UID: \"c650b71d-b6e9-4323-b8e0-52e335852881\") " pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.683095 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/aea84e78-a09c-400c-9e99-d5e7aca0ce51-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-tdzdw\" (UID: \"aea84e78-a09c-400c-9e99-d5e7aca0ce51\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.683139 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/aea84e78-a09c-400c-9e99-d5e7aca0ce51-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-tdzdw\" (UID: \"aea84e78-a09c-400c-9e99-d5e7aca0ce51\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.683171 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/f2bcdf28-cd59-4f1f-b55d-614395ff2788-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-hjnrk\" (UID: \"f2bcdf28-cd59-4f1f-b55d-614395ff2788\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hjnrk" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.683206 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/aea84e78-a09c-400c-9e99-d5e7aca0ce51-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-tdzdw\" (UID: \"aea84e78-a09c-400c-9e99-d5e7aca0ce51\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" Dec 09 09:52:56 crc kubenswrapper[4824]: E1209 09:52:56.683796 4824 secret.go:188] Couldn't get secret openshift-monitoring/openshift-state-metrics-tls: secret "openshift-state-metrics-tls" not found Dec 09 09:52:56 crc kubenswrapper[4824]: E1209 09:52:56.683913 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f2bcdf28-cd59-4f1f-b55d-614395ff2788-openshift-state-metrics-tls podName:f2bcdf28-cd59-4f1f-b55d-614395ff2788 nodeName:}" failed. No retries permitted until 2025-12-09 09:52:57.183878891 +0000 UTC m=+333.518383718 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "openshift-state-metrics-tls" (UniqueName: "kubernetes.io/secret/f2bcdf28-cd59-4f1f-b55d-614395ff2788-openshift-state-metrics-tls") pod "openshift-state-metrics-566fddb674-hjnrk" (UID: "f2bcdf28-cd59-4f1f-b55d-614395ff2788") : secret "openshift-state-metrics-tls" not found Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.685194 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/f2bcdf28-cd59-4f1f-b55d-614395ff2788-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-hjnrk\" (UID: \"f2bcdf28-cd59-4f1f-b55d-614395ff2788\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hjnrk" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.701856 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/f2bcdf28-cd59-4f1f-b55d-614395ff2788-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-hjnrk\" (UID: \"f2bcdf28-cd59-4f1f-b55d-614395ff2788\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hjnrk" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.706770 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8w2d\" (UniqueName: \"kubernetes.io/projected/f2bcdf28-cd59-4f1f-b55d-614395ff2788-kube-api-access-t8w2d\") pod \"openshift-state-metrics-566fddb674-hjnrk\" (UID: \"f2bcdf28-cd59-4f1f-b55d-614395ff2788\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hjnrk" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.784845 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/aea84e78-a09c-400c-9e99-d5e7aca0ce51-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-tdzdw\" (UID: \"aea84e78-a09c-400c-9e99-d5e7aca0ce51\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.784951 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/aea84e78-a09c-400c-9e99-d5e7aca0ce51-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-tdzdw\" (UID: \"aea84e78-a09c-400c-9e99-d5e7aca0ce51\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.785017 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c650b71d-b6e9-4323-b8e0-52e335852881-sys\") pod \"node-exporter-78nq2\" (UID: \"c650b71d-b6e9-4323-b8e0-52e335852881\") " pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.785047 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/c650b71d-b6e9-4323-b8e0-52e335852881-node-exporter-textfile\") pod \"node-exporter-78nq2\" (UID: \"c650b71d-b6e9-4323-b8e0-52e335852881\") " pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.785094 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/c650b71d-b6e9-4323-b8e0-52e335852881-root\") pod \"node-exporter-78nq2\" (UID: \"c650b71d-b6e9-4323-b8e0-52e335852881\") " pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.785126 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/c650b71d-b6e9-4323-b8e0-52e335852881-node-exporter-wtmp\") pod \"node-exporter-78nq2\" (UID: \"c650b71d-b6e9-4323-b8e0-52e335852881\") " pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.785207 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/aea84e78-a09c-400c-9e99-d5e7aca0ce51-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-tdzdw\" (UID: \"aea84e78-a09c-400c-9e99-d5e7aca0ce51\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.785236 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/c650b71d-b6e9-4323-b8e0-52e335852881-metrics-client-ca\") pod \"node-exporter-78nq2\" (UID: \"c650b71d-b6e9-4323-b8e0-52e335852881\") " pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.785278 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/c650b71d-b6e9-4323-b8e0-52e335852881-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-78nq2\" (UID: \"c650b71d-b6e9-4323-b8e0-52e335852881\") " pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.785307 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/aea84e78-a09c-400c-9e99-d5e7aca0ce51-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-tdzdw\" (UID: \"aea84e78-a09c-400c-9e99-d5e7aca0ce51\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.785339 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcd94\" (UniqueName: \"kubernetes.io/projected/aea84e78-a09c-400c-9e99-d5e7aca0ce51-kube-api-access-lcd94\") pod \"kube-state-metrics-777cb5bd5d-tdzdw\" (UID: \"aea84e78-a09c-400c-9e99-d5e7aca0ce51\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.785368 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77gln\" (UniqueName: \"kubernetes.io/projected/c650b71d-b6e9-4323-b8e0-52e335852881-kube-api-access-77gln\") pod \"node-exporter-78nq2\" (UID: \"c650b71d-b6e9-4323-b8e0-52e335852881\") " pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.785393 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/c650b71d-b6e9-4323-b8e0-52e335852881-node-exporter-tls\") pod \"node-exporter-78nq2\" (UID: \"c650b71d-b6e9-4323-b8e0-52e335852881\") " pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.785420 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/aea84e78-a09c-400c-9e99-d5e7aca0ce51-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-tdzdw\" (UID: \"aea84e78-a09c-400c-9e99-d5e7aca0ce51\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.785885 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/c650b71d-b6e9-4323-b8e0-52e335852881-node-exporter-textfile\") pod \"node-exporter-78nq2\" (UID: \"c650b71d-b6e9-4323-b8e0-52e335852881\") " pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.786256 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/c650b71d-b6e9-4323-b8e0-52e335852881-node-exporter-wtmp\") pod \"node-exporter-78nq2\" (UID: \"c650b71d-b6e9-4323-b8e0-52e335852881\") " pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.786271 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/aea84e78-a09c-400c-9e99-d5e7aca0ce51-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-tdzdw\" (UID: \"aea84e78-a09c-400c-9e99-d5e7aca0ce51\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.786308 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/c650b71d-b6e9-4323-b8e0-52e335852881-root\") pod \"node-exporter-78nq2\" (UID: \"c650b71d-b6e9-4323-b8e0-52e335852881\") " pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.786502 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/aea84e78-a09c-400c-9e99-d5e7aca0ce51-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-tdzdw\" (UID: \"aea84e78-a09c-400c-9e99-d5e7aca0ce51\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.786674 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/aea84e78-a09c-400c-9e99-d5e7aca0ce51-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-tdzdw\" (UID: \"aea84e78-a09c-400c-9e99-d5e7aca0ce51\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.787275 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c650b71d-b6e9-4323-b8e0-52e335852881-sys\") pod \"node-exporter-78nq2\" (UID: \"c650b71d-b6e9-4323-b8e0-52e335852881\") " pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.787591 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/c650b71d-b6e9-4323-b8e0-52e335852881-metrics-client-ca\") pod \"node-exporter-78nq2\" (UID: \"c650b71d-b6e9-4323-b8e0-52e335852881\") " pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.791775 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/aea84e78-a09c-400c-9e99-d5e7aca0ce51-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-tdzdw\" (UID: \"aea84e78-a09c-400c-9e99-d5e7aca0ce51\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.796475 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/c650b71d-b6e9-4323-b8e0-52e335852881-node-exporter-tls\") pod \"node-exporter-78nq2\" (UID: \"c650b71d-b6e9-4323-b8e0-52e335852881\") " pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.797323 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/c650b71d-b6e9-4323-b8e0-52e335852881-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-78nq2\" (UID: \"c650b71d-b6e9-4323-b8e0-52e335852881\") " pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.801554 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/aea84e78-a09c-400c-9e99-d5e7aca0ce51-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-tdzdw\" (UID: \"aea84e78-a09c-400c-9e99-d5e7aca0ce51\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.810442 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcd94\" (UniqueName: \"kubernetes.io/projected/aea84e78-a09c-400c-9e99-d5e7aca0ce51-kube-api-access-lcd94\") pod \"kube-state-metrics-777cb5bd5d-tdzdw\" (UID: \"aea84e78-a09c-400c-9e99-d5e7aca0ce51\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.812545 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77gln\" (UniqueName: \"kubernetes.io/projected/c650b71d-b6e9-4323-b8e0-52e335852881-kube-api-access-77gln\") pod \"node-exporter-78nq2\" (UID: \"c650b71d-b6e9-4323-b8e0-52e335852881\") " pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.883448 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-78nq2" Dec 09 09:52:56 crc kubenswrapper[4824]: I1209 09:52:56.896480 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" Dec 09 09:52:57 crc kubenswrapper[4824]: I1209 09:52:57.194562 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/f2bcdf28-cd59-4f1f-b55d-614395ff2788-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-hjnrk\" (UID: \"f2bcdf28-cd59-4f1f-b55d-614395ff2788\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hjnrk" Dec 09 09:52:57 crc kubenswrapper[4824]: I1209 09:52:57.199540 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/f2bcdf28-cd59-4f1f-b55d-614395ff2788-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-hjnrk\" (UID: \"f2bcdf28-cd59-4f1f-b55d-614395ff2788\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hjnrk" Dec 09 09:52:57 crc kubenswrapper[4824]: I1209 09:52:57.399396 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-566fddb674-hjnrk" Dec 09 09:52:57 crc kubenswrapper[4824]: I1209 09:52:57.549904 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw"] Dec 09 09:52:58 crc kubenswrapper[4824]: W1209 09:52:57.660757 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaea84e78_a09c_400c_9e99_d5e7aca0ce51.slice/crio-ede0cd42e9090f604c93a512cce082b19ddb09123a92e0004a09e3e3d0c8afb0 WatchSource:0}: Error finding container ede0cd42e9090f604c93a512cce082b19ddb09123a92e0004a09e3e3d0c8afb0: Status 404 returned error can't find the container with id ede0cd42e9090f604c93a512cce082b19ddb09123a92e0004a09e3e3d0c8afb0 Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.678874 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-78nq2" event={"ID":"c650b71d-b6e9-4323-b8e0-52e335852881","Type":"ContainerStarted","Data":"9d208dfecae6181c7e120ebd4879ae3014509a23987b71ebae67dbf3afc3444e"} Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.692997 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" event={"ID":"aea84e78-a09c-400c-9e99-d5e7aca0ce51","Type":"ContainerStarted","Data":"ede0cd42e9090f604c93a512cce082b19ddb09123a92e0004a09e3e3d0c8afb0"} Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.759187 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.761559 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.767672 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls-assets-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.767932 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-generated" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.768092 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.768237 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-metric" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.768377 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.768544 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-web-config" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.768686 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-web" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.802728 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"alertmanager-trusted-ca-bundle" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.804084 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-dockercfg-hq6bk" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.812070 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.930029 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/26c6a6ba-64bf-4473-816d-e9ee424bf97b-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.930123 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/26c6a6ba-64bf-4473-816d-e9ee424bf97b-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.930164 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/26c6a6ba-64bf-4473-816d-e9ee424bf97b-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.930191 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/26c6a6ba-64bf-4473-816d-e9ee424bf97b-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.930214 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/26c6a6ba-64bf-4473-816d-e9ee424bf97b-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.930237 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vt862\" (UniqueName: \"kubernetes.io/projected/26c6a6ba-64bf-4473-816d-e9ee424bf97b-kube-api-access-vt862\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.930262 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/26c6a6ba-64bf-4473-816d-e9ee424bf97b-web-config\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.930285 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/26c6a6ba-64bf-4473-816d-e9ee424bf97b-tls-assets\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.930319 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/26c6a6ba-64bf-4473-816d-e9ee424bf97b-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.930351 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/26c6a6ba-64bf-4473-816d-e9ee424bf97b-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.930394 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/26c6a6ba-64bf-4473-816d-e9ee424bf97b-config-out\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:57.930460 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/26c6a6ba-64bf-4473-816d-e9ee424bf97b-config-volume\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.031268 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/26c6a6ba-64bf-4473-816d-e9ee424bf97b-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.031370 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/26c6a6ba-64bf-4473-816d-e9ee424bf97b-config-out\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.031429 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/26c6a6ba-64bf-4473-816d-e9ee424bf97b-config-volume\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.031462 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/26c6a6ba-64bf-4473-816d-e9ee424bf97b-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.031485 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/26c6a6ba-64bf-4473-816d-e9ee424bf97b-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.031508 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/26c6a6ba-64bf-4473-816d-e9ee424bf97b-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.031531 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/26c6a6ba-64bf-4473-816d-e9ee424bf97b-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.031550 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/26c6a6ba-64bf-4473-816d-e9ee424bf97b-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.031571 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/26c6a6ba-64bf-4473-816d-e9ee424bf97b-web-config\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.031590 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vt862\" (UniqueName: \"kubernetes.io/projected/26c6a6ba-64bf-4473-816d-e9ee424bf97b-kube-api-access-vt862\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.031612 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/26c6a6ba-64bf-4473-816d-e9ee424bf97b-tls-assets\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.031638 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/26c6a6ba-64bf-4473-816d-e9ee424bf97b-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.032082 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/26c6a6ba-64bf-4473-816d-e9ee424bf97b-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.039057 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/26c6a6ba-64bf-4473-816d-e9ee424bf97b-config-volume\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.040204 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/26c6a6ba-64bf-4473-816d-e9ee424bf97b-web-config\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: E1209 09:52:58.040341 4824 secret.go:188] Couldn't get secret openshift-monitoring/alertmanager-main-tls: secret "alertmanager-main-tls" not found Dec 09 09:52:58 crc kubenswrapper[4824]: E1209 09:52:58.040417 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/26c6a6ba-64bf-4473-816d-e9ee424bf97b-secret-alertmanager-main-tls podName:26c6a6ba-64bf-4473-816d-e9ee424bf97b nodeName:}" failed. No retries permitted until 2025-12-09 09:52:58.540389756 +0000 UTC m=+334.874894423 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-alertmanager-main-tls" (UniqueName: "kubernetes.io/secret/26c6a6ba-64bf-4473-816d-e9ee424bf97b-secret-alertmanager-main-tls") pod "alertmanager-main-0" (UID: "26c6a6ba-64bf-4473-816d-e9ee424bf97b") : secret "alertmanager-main-tls" not found Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.040803 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/26c6a6ba-64bf-4473-816d-e9ee424bf97b-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.041526 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/26c6a6ba-64bf-4473-816d-e9ee424bf97b-config-out\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.043174 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/26c6a6ba-64bf-4473-816d-e9ee424bf97b-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.045891 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/26c6a6ba-64bf-4473-816d-e9ee424bf97b-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.051347 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/26c6a6ba-64bf-4473-816d-e9ee424bf97b-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.058682 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/26c6a6ba-64bf-4473-816d-e9ee424bf97b-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.063214 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/26c6a6ba-64bf-4473-816d-e9ee424bf97b-tls-assets\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.064056 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vt862\" (UniqueName: \"kubernetes.io/projected/26c6a6ba-64bf-4473-816d-e9ee424bf97b-kube-api-access-vt862\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.641462 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/26c6a6ba-64bf-4473-816d-e9ee424bf97b-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.656432 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/26c6a6ba-64bf-4473-816d-e9ee424bf97b-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"26c6a6ba-64bf-4473-816d-e9ee424bf97b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.671483 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-hjnrk"] Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.751267 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f"] Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.754005 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.762389 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-dockercfg-g46f7" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.762649 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-web" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.762830 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.763006 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-metrics" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.763196 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-grpc-tls-f6c5h5nvib2t4" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.763196 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-rules" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.763308 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-tls" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.781892 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.799558 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f"] Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.950346 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/d030e6a9-d47a-47bf-9c24-ff0ca58f71e6-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-7fc6d7f97-nsq7f\" (UID: \"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6\") " pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.950539 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d030e6a9-d47a-47bf-9c24-ff0ca58f71e6-metrics-client-ca\") pod \"thanos-querier-7fc6d7f97-nsq7f\" (UID: \"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6\") " pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.950667 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/d030e6a9-d47a-47bf-9c24-ff0ca58f71e6-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-7fc6d7f97-nsq7f\" (UID: \"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6\") " pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.951815 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/d030e6a9-d47a-47bf-9c24-ff0ca58f71e6-secret-thanos-querier-tls\") pod \"thanos-querier-7fc6d7f97-nsq7f\" (UID: \"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6\") " pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.951880 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/d030e6a9-d47a-47bf-9c24-ff0ca58f71e6-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-7fc6d7f97-nsq7f\" (UID: \"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6\") " pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.951968 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrtwk\" (UniqueName: \"kubernetes.io/projected/d030e6a9-d47a-47bf-9c24-ff0ca58f71e6-kube-api-access-zrtwk\") pod \"thanos-querier-7fc6d7f97-nsq7f\" (UID: \"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6\") " pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.952036 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/d030e6a9-d47a-47bf-9c24-ff0ca58f71e6-secret-grpc-tls\") pod \"thanos-querier-7fc6d7f97-nsq7f\" (UID: \"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6\") " pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:58 crc kubenswrapper[4824]: I1209 09:52:58.952082 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/d030e6a9-d47a-47bf-9c24-ff0ca58f71e6-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-7fc6d7f97-nsq7f\" (UID: \"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6\") " pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.055737 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrtwk\" (UniqueName: \"kubernetes.io/projected/d030e6a9-d47a-47bf-9c24-ff0ca58f71e6-kube-api-access-zrtwk\") pod \"thanos-querier-7fc6d7f97-nsq7f\" (UID: \"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6\") " pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.057641 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/d030e6a9-d47a-47bf-9c24-ff0ca58f71e6-secret-grpc-tls\") pod \"thanos-querier-7fc6d7f97-nsq7f\" (UID: \"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6\") " pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.057738 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/d030e6a9-d47a-47bf-9c24-ff0ca58f71e6-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-7fc6d7f97-nsq7f\" (UID: \"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6\") " pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.057848 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/d030e6a9-d47a-47bf-9c24-ff0ca58f71e6-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-7fc6d7f97-nsq7f\" (UID: \"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6\") " pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.057903 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d030e6a9-d47a-47bf-9c24-ff0ca58f71e6-metrics-client-ca\") pod \"thanos-querier-7fc6d7f97-nsq7f\" (UID: \"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6\") " pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.062652 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/d030e6a9-d47a-47bf-9c24-ff0ca58f71e6-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-7fc6d7f97-nsq7f\" (UID: \"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6\") " pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.062809 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/d030e6a9-d47a-47bf-9c24-ff0ca58f71e6-secret-thanos-querier-tls\") pod \"thanos-querier-7fc6d7f97-nsq7f\" (UID: \"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6\") " pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.062841 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/d030e6a9-d47a-47bf-9c24-ff0ca58f71e6-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-7fc6d7f97-nsq7f\" (UID: \"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6\") " pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.069072 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d030e6a9-d47a-47bf-9c24-ff0ca58f71e6-metrics-client-ca\") pod \"thanos-querier-7fc6d7f97-nsq7f\" (UID: \"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6\") " pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.085052 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/d030e6a9-d47a-47bf-9c24-ff0ca58f71e6-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-7fc6d7f97-nsq7f\" (UID: \"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6\") " pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.088347 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/d030e6a9-d47a-47bf-9c24-ff0ca58f71e6-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-7fc6d7f97-nsq7f\" (UID: \"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6\") " pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.088676 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/d030e6a9-d47a-47bf-9c24-ff0ca58f71e6-secret-thanos-querier-tls\") pod \"thanos-querier-7fc6d7f97-nsq7f\" (UID: \"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6\") " pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.090159 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/d030e6a9-d47a-47bf-9c24-ff0ca58f71e6-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-7fc6d7f97-nsq7f\" (UID: \"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6\") " pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.098257 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/d030e6a9-d47a-47bf-9c24-ff0ca58f71e6-secret-grpc-tls\") pod \"thanos-querier-7fc6d7f97-nsq7f\" (UID: \"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6\") " pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.104392 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/d030e6a9-d47a-47bf-9c24-ff0ca58f71e6-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-7fc6d7f97-nsq7f\" (UID: \"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6\") " pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.123306 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrtwk\" (UniqueName: \"kubernetes.io/projected/d030e6a9-d47a-47bf-9c24-ff0ca58f71e6-kube-api-access-zrtwk\") pod \"thanos-querier-7fc6d7f97-nsq7f\" (UID: \"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6\") " pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.385470 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.477461 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Dec 09 09:52:59 crc kubenswrapper[4824]: W1209 09:52:59.488543 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26c6a6ba_64bf_4473_816d_e9ee424bf97b.slice/crio-9d7b3432fd4728541e93dd0187789592f428ff22437ba9baed36c2f1899291ec WatchSource:0}: Error finding container 9d7b3432fd4728541e93dd0187789592f428ff22437ba9baed36c2f1899291ec: Status 404 returned error can't find the container with id 9d7b3432fd4728541e93dd0187789592f428ff22437ba9baed36c2f1899291ec Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.721686 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-hjnrk" event={"ID":"f2bcdf28-cd59-4f1f-b55d-614395ff2788","Type":"ContainerStarted","Data":"c228e8ec53f0a5fd2d4e6c9d7ab326eb9cac73a1fae6f1745455ea05ec3aa980"} Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.721762 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-hjnrk" event={"ID":"f2bcdf28-cd59-4f1f-b55d-614395ff2788","Type":"ContainerStarted","Data":"732537b4c7379bd991e163fee5298c8d9cf449df2afc2af54f5701c304f55f75"} Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.721792 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-hjnrk" event={"ID":"f2bcdf28-cd59-4f1f-b55d-614395ff2788","Type":"ContainerStarted","Data":"a4c9b73c601c75a06f8a3f115562a471feb4e8662cd2d8deece7e44733369aa5"} Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.722637 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"26c6a6ba-64bf-4473-816d-e9ee424bf97b","Type":"ContainerStarted","Data":"9d7b3432fd4728541e93dd0187789592f428ff22437ba9baed36c2f1899291ec"} Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.723797 4824 generic.go:334] "Generic (PLEG): container finished" podID="c650b71d-b6e9-4323-b8e0-52e335852881" containerID="da21186488c2f1c1b2d8537b9d96ff8b8176ec8591c96908920d115682f78428" exitCode=0 Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.723835 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-78nq2" event={"ID":"c650b71d-b6e9-4323-b8e0-52e335852881","Type":"ContainerDied","Data":"da21186488c2f1c1b2d8537b9d96ff8b8176ec8591c96908920d115682f78428"} Dec 09 09:52:59 crc kubenswrapper[4824]: I1209 09:52:59.877749 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f"] Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.053432 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-jq6bp"] Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.054794 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.079133 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-jq6bp"] Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.187041 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ffa85659-bcb5-43c8-9270-ab07d6ec9b70-installation-pull-secrets\") pod \"image-registry-66df7c8f76-jq6bp\" (UID: \"ffa85659-bcb5-43c8-9270-ab07d6ec9b70\") " pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.187110 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ffa85659-bcb5-43c8-9270-ab07d6ec9b70-ca-trust-extracted\") pod \"image-registry-66df7c8f76-jq6bp\" (UID: \"ffa85659-bcb5-43c8-9270-ab07d6ec9b70\") " pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.187146 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ffa85659-bcb5-43c8-9270-ab07d6ec9b70-bound-sa-token\") pod \"image-registry-66df7c8f76-jq6bp\" (UID: \"ffa85659-bcb5-43c8-9270-ab07d6ec9b70\") " pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.187175 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2dl4\" (UniqueName: \"kubernetes.io/projected/ffa85659-bcb5-43c8-9270-ab07d6ec9b70-kube-api-access-w2dl4\") pod \"image-registry-66df7c8f76-jq6bp\" (UID: \"ffa85659-bcb5-43c8-9270-ab07d6ec9b70\") " pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.187209 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ffa85659-bcb5-43c8-9270-ab07d6ec9b70-registry-tls\") pod \"image-registry-66df7c8f76-jq6bp\" (UID: \"ffa85659-bcb5-43c8-9270-ab07d6ec9b70\") " pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.187279 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-jq6bp\" (UID: \"ffa85659-bcb5-43c8-9270-ab07d6ec9b70\") " pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.187320 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ffa85659-bcb5-43c8-9270-ab07d6ec9b70-registry-certificates\") pod \"image-registry-66df7c8f76-jq6bp\" (UID: \"ffa85659-bcb5-43c8-9270-ab07d6ec9b70\") " pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.187352 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ffa85659-bcb5-43c8-9270-ab07d6ec9b70-trusted-ca\") pod \"image-registry-66df7c8f76-jq6bp\" (UID: \"ffa85659-bcb5-43c8-9270-ab07d6ec9b70\") " pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.216005 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-jq6bp\" (UID: \"ffa85659-bcb5-43c8-9270-ab07d6ec9b70\") " pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.290325 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ffa85659-bcb5-43c8-9270-ab07d6ec9b70-ca-trust-extracted\") pod \"image-registry-66df7c8f76-jq6bp\" (UID: \"ffa85659-bcb5-43c8-9270-ab07d6ec9b70\") " pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.290441 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ffa85659-bcb5-43c8-9270-ab07d6ec9b70-bound-sa-token\") pod \"image-registry-66df7c8f76-jq6bp\" (UID: \"ffa85659-bcb5-43c8-9270-ab07d6ec9b70\") " pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.290501 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2dl4\" (UniqueName: \"kubernetes.io/projected/ffa85659-bcb5-43c8-9270-ab07d6ec9b70-kube-api-access-w2dl4\") pod \"image-registry-66df7c8f76-jq6bp\" (UID: \"ffa85659-bcb5-43c8-9270-ab07d6ec9b70\") " pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.290627 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ffa85659-bcb5-43c8-9270-ab07d6ec9b70-registry-tls\") pod \"image-registry-66df7c8f76-jq6bp\" (UID: \"ffa85659-bcb5-43c8-9270-ab07d6ec9b70\") " pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.290728 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ffa85659-bcb5-43c8-9270-ab07d6ec9b70-registry-certificates\") pod \"image-registry-66df7c8f76-jq6bp\" (UID: \"ffa85659-bcb5-43c8-9270-ab07d6ec9b70\") " pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.290773 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ffa85659-bcb5-43c8-9270-ab07d6ec9b70-trusted-ca\") pod \"image-registry-66df7c8f76-jq6bp\" (UID: \"ffa85659-bcb5-43c8-9270-ab07d6ec9b70\") " pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.290851 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ffa85659-bcb5-43c8-9270-ab07d6ec9b70-installation-pull-secrets\") pod \"image-registry-66df7c8f76-jq6bp\" (UID: \"ffa85659-bcb5-43c8-9270-ab07d6ec9b70\") " pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.291966 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ffa85659-bcb5-43c8-9270-ab07d6ec9b70-ca-trust-extracted\") pod \"image-registry-66df7c8f76-jq6bp\" (UID: \"ffa85659-bcb5-43c8-9270-ab07d6ec9b70\") " pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.294153 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ffa85659-bcb5-43c8-9270-ab07d6ec9b70-registry-certificates\") pod \"image-registry-66df7c8f76-jq6bp\" (UID: \"ffa85659-bcb5-43c8-9270-ab07d6ec9b70\") " pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.294317 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ffa85659-bcb5-43c8-9270-ab07d6ec9b70-trusted-ca\") pod \"image-registry-66df7c8f76-jq6bp\" (UID: \"ffa85659-bcb5-43c8-9270-ab07d6ec9b70\") " pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.298366 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ffa85659-bcb5-43c8-9270-ab07d6ec9b70-registry-tls\") pod \"image-registry-66df7c8f76-jq6bp\" (UID: \"ffa85659-bcb5-43c8-9270-ab07d6ec9b70\") " pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.304353 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ffa85659-bcb5-43c8-9270-ab07d6ec9b70-installation-pull-secrets\") pod \"image-registry-66df7c8f76-jq6bp\" (UID: \"ffa85659-bcb5-43c8-9270-ab07d6ec9b70\") " pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.318084 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2dl4\" (UniqueName: \"kubernetes.io/projected/ffa85659-bcb5-43c8-9270-ab07d6ec9b70-kube-api-access-w2dl4\") pod \"image-registry-66df7c8f76-jq6bp\" (UID: \"ffa85659-bcb5-43c8-9270-ab07d6ec9b70\") " pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.329843 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ffa85659-bcb5-43c8-9270-ab07d6ec9b70-bound-sa-token\") pod \"image-registry-66df7c8f76-jq6bp\" (UID: \"ffa85659-bcb5-43c8-9270-ab07d6ec9b70\") " pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.385567 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.739140 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" event={"ID":"aea84e78-a09c-400c-9e99-d5e7aca0ce51","Type":"ContainerStarted","Data":"5f256092b78fc02ea0a22979ad55f5eba0b453417fc7dbdda39eb62923f12cf4"} Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.739599 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" event={"ID":"aea84e78-a09c-400c-9e99-d5e7aca0ce51","Type":"ContainerStarted","Data":"33df6497dcc50491adfec923118892faa0ae9b820f6e623481f3cb63638f82db"} Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.742235 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-78nq2" event={"ID":"c650b71d-b6e9-4323-b8e0-52e335852881","Type":"ContainerStarted","Data":"0481aa0537290a365932b050d0868075e0bbb507ecb4667f22cc348a932de075"} Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.742271 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-78nq2" event={"ID":"c650b71d-b6e9-4323-b8e0-52e335852881","Type":"ContainerStarted","Data":"dc44b0671ff5f5d3c6489775de2e80c4be0ba9648603c75b5b57c231a0cd259a"} Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.746145 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" event={"ID":"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6","Type":"ContainerStarted","Data":"05667d6c339456fdf6e595a44a08a9ba632b8aa9af7d5e6afa9a8e2b18cc7aef"} Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.958440 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-78nq2" podStartSLOduration=2.895771113 podStartE2EDuration="4.958377016s" podCreationTimestamp="2025-12-09 09:52:56 +0000 UTC" firstStartedPulling="2025-12-09 09:52:56.940536026 +0000 UTC m=+333.275040693" lastFinishedPulling="2025-12-09 09:52:59.003141939 +0000 UTC m=+335.337646596" observedRunningTime="2025-12-09 09:53:00.775918113 +0000 UTC m=+337.110422800" watchObservedRunningTime="2025-12-09 09:53:00.958377016 +0000 UTC m=+337.292881683" Dec 09 09:53:00 crc kubenswrapper[4824]: I1209 09:53:00.990869 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-jq6bp"] Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.310334 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-684fc64449-59s5g"] Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.312264 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.346269 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-684fc64449-59s5g"] Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.352903 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/536256e5-73d0-4de7-8f6d-4e5d449b84ef-oauth-serving-cert\") pod \"console-684fc64449-59s5g\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.352965 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/536256e5-73d0-4de7-8f6d-4e5d449b84ef-console-serving-cert\") pod \"console-684fc64449-59s5g\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.352993 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c22f\" (UniqueName: \"kubernetes.io/projected/536256e5-73d0-4de7-8f6d-4e5d449b84ef-kube-api-access-9c22f\") pod \"console-684fc64449-59s5g\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.353035 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/536256e5-73d0-4de7-8f6d-4e5d449b84ef-console-oauth-config\") pod \"console-684fc64449-59s5g\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.353268 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/536256e5-73d0-4de7-8f6d-4e5d449b84ef-trusted-ca-bundle\") pod \"console-684fc64449-59s5g\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.353374 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/536256e5-73d0-4de7-8f6d-4e5d449b84ef-console-config\") pod \"console-684fc64449-59s5g\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.353428 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/536256e5-73d0-4de7-8f6d-4e5d449b84ef-service-ca\") pod \"console-684fc64449-59s5g\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.455276 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/536256e5-73d0-4de7-8f6d-4e5d449b84ef-console-config\") pod \"console-684fc64449-59s5g\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.455339 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/536256e5-73d0-4de7-8f6d-4e5d449b84ef-service-ca\") pod \"console-684fc64449-59s5g\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.455379 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/536256e5-73d0-4de7-8f6d-4e5d449b84ef-oauth-serving-cert\") pod \"console-684fc64449-59s5g\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.455412 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/536256e5-73d0-4de7-8f6d-4e5d449b84ef-console-serving-cert\") pod \"console-684fc64449-59s5g\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.455440 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c22f\" (UniqueName: \"kubernetes.io/projected/536256e5-73d0-4de7-8f6d-4e5d449b84ef-kube-api-access-9c22f\") pod \"console-684fc64449-59s5g\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.455501 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/536256e5-73d0-4de7-8f6d-4e5d449b84ef-console-oauth-config\") pod \"console-684fc64449-59s5g\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.455549 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/536256e5-73d0-4de7-8f6d-4e5d449b84ef-trusted-ca-bundle\") pod \"console-684fc64449-59s5g\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.456818 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/536256e5-73d0-4de7-8f6d-4e5d449b84ef-trusted-ca-bundle\") pod \"console-684fc64449-59s5g\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.459437 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/536256e5-73d0-4de7-8f6d-4e5d449b84ef-console-config\") pod \"console-684fc64449-59s5g\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.460142 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/536256e5-73d0-4de7-8f6d-4e5d449b84ef-service-ca\") pod \"console-684fc64449-59s5g\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.460668 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/536256e5-73d0-4de7-8f6d-4e5d449b84ef-oauth-serving-cert\") pod \"console-684fc64449-59s5g\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.470473 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/536256e5-73d0-4de7-8f6d-4e5d449b84ef-console-oauth-config\") pod \"console-684fc64449-59s5g\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.482713 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/536256e5-73d0-4de7-8f6d-4e5d449b84ef-console-serving-cert\") pod \"console-684fc64449-59s5g\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.482886 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c22f\" (UniqueName: \"kubernetes.io/projected/536256e5-73d0-4de7-8f6d-4e5d449b84ef-kube-api-access-9c22f\") pod \"console-684fc64449-59s5g\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.632086 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.758665 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" event={"ID":"ffa85659-bcb5-43c8-9270-ab07d6ec9b70","Type":"ContainerStarted","Data":"3fc41389f6d7ba2f6ea21c45445d63b89ac13db7ecaa7589d1bbaa77f51f5fae"} Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.758737 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" event={"ID":"ffa85659-bcb5-43c8-9270-ab07d6ec9b70","Type":"ContainerStarted","Data":"a3950956bfae107ce7136a76415e8821e4a6c2f1eef6e4fd98ee42f804731152"} Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.760210 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.769006 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" event={"ID":"aea84e78-a09c-400c-9e99-d5e7aca0ce51","Type":"ContainerStarted","Data":"3db341809f34a086026a0eee9634292431423390e5e08890ed7b1b7f23849285"} Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.787527 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" podStartSLOduration=1.787499749 podStartE2EDuration="1.787499749s" podCreationTimestamp="2025-12-09 09:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:53:01.7840208 +0000 UTC m=+338.118525487" watchObservedRunningTime="2025-12-09 09:53:01.787499749 +0000 UTC m=+338.122004416" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.812201 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tdzdw" podStartSLOduration=3.25493891 podStartE2EDuration="5.812170974s" podCreationTimestamp="2025-12-09 09:52:56 +0000 UTC" firstStartedPulling="2025-12-09 09:52:57.67197148 +0000 UTC m=+334.006476147" lastFinishedPulling="2025-12-09 09:53:00.229203544 +0000 UTC m=+336.563708211" observedRunningTime="2025-12-09 09:53:01.809409938 +0000 UTC m=+338.143914625" watchObservedRunningTime="2025-12-09 09:53:01.812170974 +0000 UTC m=+338.146675651" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.940224 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/metrics-server-6448d488d4-xnf9s"] Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.941281 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.944084 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-client-certs" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.944151 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kubelet-serving-ca-bundle" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.944900 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-server-audit-profiles" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.945381 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-dockercfg-hq42b" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.945870 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-tls" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.945889 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-c9ljgokrb377b" Dec 09 09:53:01 crc kubenswrapper[4824]: I1209 09:53:01.958015 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-6448d488d4-xnf9s"] Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.083793 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/9aa0314b-a177-4589-bc32-f9caf6f8f724-secret-metrics-server-tls\") pod \"metrics-server-6448d488d4-xnf9s\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.084297 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9aa0314b-a177-4589-bc32-f9caf6f8f724-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-6448d488d4-xnf9s\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.084440 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/9aa0314b-a177-4589-bc32-f9caf6f8f724-audit-log\") pod \"metrics-server-6448d488d4-xnf9s\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.084539 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/9aa0314b-a177-4589-bc32-f9caf6f8f724-secret-metrics-client-certs\") pod \"metrics-server-6448d488d4-xnf9s\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.084670 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/9aa0314b-a177-4589-bc32-f9caf6f8f724-metrics-server-audit-profiles\") pod \"metrics-server-6448d488d4-xnf9s\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.084764 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa0314b-a177-4589-bc32-f9caf6f8f724-client-ca-bundle\") pod \"metrics-server-6448d488d4-xnf9s\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.084884 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfprf\" (UniqueName: \"kubernetes.io/projected/9aa0314b-a177-4589-bc32-f9caf6f8f724-kube-api-access-gfprf\") pod \"metrics-server-6448d488d4-xnf9s\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.186955 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/9aa0314b-a177-4589-bc32-f9caf6f8f724-secret-metrics-server-tls\") pod \"metrics-server-6448d488d4-xnf9s\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.187431 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9aa0314b-a177-4589-bc32-f9caf6f8f724-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-6448d488d4-xnf9s\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.187509 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/9aa0314b-a177-4589-bc32-f9caf6f8f724-audit-log\") pod \"metrics-server-6448d488d4-xnf9s\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.187534 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/9aa0314b-a177-4589-bc32-f9caf6f8f724-secret-metrics-client-certs\") pod \"metrics-server-6448d488d4-xnf9s\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.187564 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/9aa0314b-a177-4589-bc32-f9caf6f8f724-metrics-server-audit-profiles\") pod \"metrics-server-6448d488d4-xnf9s\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.187587 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa0314b-a177-4589-bc32-f9caf6f8f724-client-ca-bundle\") pod \"metrics-server-6448d488d4-xnf9s\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.187621 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfprf\" (UniqueName: \"kubernetes.io/projected/9aa0314b-a177-4589-bc32-f9caf6f8f724-kube-api-access-gfprf\") pod \"metrics-server-6448d488d4-xnf9s\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.188712 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9aa0314b-a177-4589-bc32-f9caf6f8f724-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-6448d488d4-xnf9s\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.189547 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/9aa0314b-a177-4589-bc32-f9caf6f8f724-audit-log\") pod \"metrics-server-6448d488d4-xnf9s\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.189867 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/9aa0314b-a177-4589-bc32-f9caf6f8f724-metrics-server-audit-profiles\") pod \"metrics-server-6448d488d4-xnf9s\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.193691 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/9aa0314b-a177-4589-bc32-f9caf6f8f724-secret-metrics-client-certs\") pod \"metrics-server-6448d488d4-xnf9s\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.194383 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa0314b-a177-4589-bc32-f9caf6f8f724-client-ca-bundle\") pod \"metrics-server-6448d488d4-xnf9s\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.195714 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/9aa0314b-a177-4589-bc32-f9caf6f8f724-secret-metrics-server-tls\") pod \"metrics-server-6448d488d4-xnf9s\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.206477 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfprf\" (UniqueName: \"kubernetes.io/projected/9aa0314b-a177-4589-bc32-f9caf6f8f724-kube-api-access-gfprf\") pod \"metrics-server-6448d488d4-xnf9s\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.266303 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.914601 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.936704 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.943406 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.943834 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-sidecar-tls" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.944056 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-grpc-tls-4ctp1219j2etj" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.944231 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-kube-rbac-proxy-web" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.945619 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.963639 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-prometheus-http-client-file" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.963751 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"serving-certs-ca-bundle" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.963959 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-web-config" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.964133 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-rbac-proxy" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.964282 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls-assets-0" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.964135 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.965092 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-k8s-rulefiles-0" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.965223 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-dockercfg-mg5cq" Dec 09 09:53:02 crc kubenswrapper[4824]: I1209 09:53:02.991839 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-trusted-ca-bundle" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.110148 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.110225 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.110256 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/844884cd-6813-4fa2-88e4-7a5994bbd3cb-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.110278 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/844884cd-6813-4fa2-88e4-7a5994bbd3cb-config-out\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.110308 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/844884cd-6813-4fa2-88e4-7a5994bbd3cb-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.110331 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-web-config\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.110372 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/844884cd-6813-4fa2-88e4-7a5994bbd3cb-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.110401 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/844884cd-6813-4fa2-88e4-7a5994bbd3cb-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.110421 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.110457 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/844884cd-6813-4fa2-88e4-7a5994bbd3cb-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.110475 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/844884cd-6813-4fa2-88e4-7a5994bbd3cb-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.110501 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-config\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.110518 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsbnf\" (UniqueName: \"kubernetes.io/projected/844884cd-6813-4fa2-88e4-7a5994bbd3cb-kube-api-access-bsbnf\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.110560 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.110587 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/844884cd-6813-4fa2-88e4-7a5994bbd3cb-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.110618 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.110651 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.110676 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.215974 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.216091 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.216165 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.216230 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.216293 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.216337 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/844884cd-6813-4fa2-88e4-7a5994bbd3cb-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.216396 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/844884cd-6813-4fa2-88e4-7a5994bbd3cb-config-out\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.216467 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/844884cd-6813-4fa2-88e4-7a5994bbd3cb-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.216502 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-web-config\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.216589 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/844884cd-6813-4fa2-88e4-7a5994bbd3cb-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.216663 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/844884cd-6813-4fa2-88e4-7a5994bbd3cb-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.216725 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.216821 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/844884cd-6813-4fa2-88e4-7a5994bbd3cb-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.216856 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/844884cd-6813-4fa2-88e4-7a5994bbd3cb-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.216922 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-config\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.216975 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsbnf\" (UniqueName: \"kubernetes.io/projected/844884cd-6813-4fa2-88e4-7a5994bbd3cb-kube-api-access-bsbnf\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.217016 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.217075 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/844884cd-6813-4fa2-88e4-7a5994bbd3cb-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.217924 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/844884cd-6813-4fa2-88e4-7a5994bbd3cb-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.220170 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/844884cd-6813-4fa2-88e4-7a5994bbd3cb-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.222600 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/844884cd-6813-4fa2-88e4-7a5994bbd3cb-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.222886 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/844884cd-6813-4fa2-88e4-7a5994bbd3cb-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.224529 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/844884cd-6813-4fa2-88e4-7a5994bbd3cb-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: E1209 09:53:03.224665 4824 configmap.go:193] Couldn't get configMap openshift-monitoring/prometheus-k8s-rulefiles-0: configmap "prometheus-k8s-rulefiles-0" not found Dec 09 09:53:03 crc kubenswrapper[4824]: E1209 09:53:03.224742 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/844884cd-6813-4fa2-88e4-7a5994bbd3cb-prometheus-k8s-rulefiles-0 podName:844884cd-6813-4fa2-88e4-7a5994bbd3cb nodeName:}" failed. No retries permitted until 2025-12-09 09:53:03.72471465 +0000 UTC m=+340.059219337 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "prometheus-k8s-rulefiles-0" (UniqueName: "kubernetes.io/configmap/844884cd-6813-4fa2-88e4-7a5994bbd3cb-prometheus-k8s-rulefiles-0") pod "prometheus-k8s-0" (UID: "844884cd-6813-4fa2-88e4-7a5994bbd3cb") : configmap "prometheus-k8s-rulefiles-0" not found Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.235545 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-config\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.238957 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/844884cd-6813-4fa2-88e4-7a5994bbd3cb-config-out\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.245234 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.247008 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.247671 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/844884cd-6813-4fa2-88e4-7a5994bbd3cb-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.250624 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.251124 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-web-config\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.259435 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.259856 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.264592 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.268943 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/844884cd-6813-4fa2-88e4-7a5994bbd3cb-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.269770 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsbnf\" (UniqueName: \"kubernetes.io/projected/844884cd-6813-4fa2-88e4-7a5994bbd3cb-kube-api-access-bsbnf\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.285510 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/monitoring-plugin-6c57467cd4-4q6wt"] Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.287083 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-6c57467cd4-4q6wt" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.294334 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"monitoring-plugin-cert" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.294604 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"default-dockercfg-6tstp" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.296776 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-6c57467cd4-4q6wt"] Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.433001 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3-monitoring-plugin-cert\") pod \"monitoring-plugin-6c57467cd4-4q6wt\" (UID: \"0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3\") " pod="openshift-monitoring/monitoring-plugin-6c57467cd4-4q6wt" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.534848 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3-monitoring-plugin-cert\") pod \"monitoring-plugin-6c57467cd4-4q6wt\" (UID: \"0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3\") " pod="openshift-monitoring/monitoring-plugin-6c57467cd4-4q6wt" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.541363 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3-monitoring-plugin-cert\") pod \"monitoring-plugin-6c57467cd4-4q6wt\" (UID: \"0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3\") " pod="openshift-monitoring/monitoring-plugin-6c57467cd4-4q6wt" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.668481 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-6c57467cd4-4q6wt" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.738650 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/844884cd-6813-4fa2-88e4-7a5994bbd3cb-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.744814 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/844884cd-6813-4fa2-88e4-7a5994bbd3cb-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"844884cd-6813-4fa2-88e4-7a5994bbd3cb\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:03 crc kubenswrapper[4824]: I1209 09:53:03.903989 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:05 crc kubenswrapper[4824]: I1209 09:53:05.234828 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-684fc64449-59s5g"] Dec 09 09:53:05 crc kubenswrapper[4824]: I1209 09:53:05.283554 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-6c57467cd4-4q6wt"] Dec 09 09:53:05 crc kubenswrapper[4824]: I1209 09:53:05.361419 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-6448d488d4-xnf9s"] Dec 09 09:53:05 crc kubenswrapper[4824]: W1209 09:53:05.378037 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9aa0314b_a177_4589_bc32_f9caf6f8f724.slice/crio-509989dd226b9819afb3d9917ad61603277049db1578fcba6fefba79ce205ab4 WatchSource:0}: Error finding container 509989dd226b9819afb3d9917ad61603277049db1578fcba6fefba79ce205ab4: Status 404 returned error can't find the container with id 509989dd226b9819afb3d9917ad61603277049db1578fcba6fefba79ce205ab4 Dec 09 09:53:05 crc kubenswrapper[4824]: I1209 09:53:05.473747 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Dec 09 09:53:05 crc kubenswrapper[4824]: I1209 09:53:05.853828 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" event={"ID":"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6","Type":"ContainerStarted","Data":"2d23440a0bbc590e190e2b83ca15028488b0b539c60b3e08ad8baf47ceb3ec59"} Dec 09 09:53:05 crc kubenswrapper[4824]: I1209 09:53:05.854322 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" event={"ID":"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6","Type":"ContainerStarted","Data":"0f89e65498081381b98846c34c57855e15dfd012c503211cf391727eb286e614"} Dec 09 09:53:05 crc kubenswrapper[4824]: I1209 09:53:05.854333 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" event={"ID":"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6","Type":"ContainerStarted","Data":"8e9907b79049ea8dc696a064611a26c057a5222cced058c7adf53b08ca9f2478"} Dec 09 09:53:05 crc kubenswrapper[4824]: I1209 09:53:05.855473 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-684fc64449-59s5g" event={"ID":"536256e5-73d0-4de7-8f6d-4e5d449b84ef","Type":"ContainerStarted","Data":"2b8b31e3c2c987595769cb00f48cfbd5cf000e9273ee2990febb6749308b007b"} Dec 09 09:53:05 crc kubenswrapper[4824]: I1209 09:53:05.855567 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-684fc64449-59s5g" event={"ID":"536256e5-73d0-4de7-8f6d-4e5d449b84ef","Type":"ContainerStarted","Data":"1bba57fbb6e3ab84e8fb5d4c7d6e01d3675150d3a55100691f2b7e50f6012582"} Dec 09 09:53:05 crc kubenswrapper[4824]: I1209 09:53:05.856967 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" event={"ID":"9aa0314b-a177-4589-bc32-f9caf6f8f724","Type":"ContainerStarted","Data":"509989dd226b9819afb3d9917ad61603277049db1578fcba6fefba79ce205ab4"} Dec 09 09:53:05 crc kubenswrapper[4824]: I1209 09:53:05.861426 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-hjnrk" event={"ID":"f2bcdf28-cd59-4f1f-b55d-614395ff2788","Type":"ContainerStarted","Data":"0aa6807b38e3e31eecec392dacff5274813f43bb14629df8a6758d4abc6dd028"} Dec 09 09:53:05 crc kubenswrapper[4824]: I1209 09:53:05.864831 4824 generic.go:334] "Generic (PLEG): container finished" podID="26c6a6ba-64bf-4473-816d-e9ee424bf97b" containerID="31d0cf5512a1cc5dd661ae6e504d43d09c1bc8e8da37b0a9967bc3e9d4d6db40" exitCode=0 Dec 09 09:53:05 crc kubenswrapper[4824]: I1209 09:53:05.864908 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"26c6a6ba-64bf-4473-816d-e9ee424bf97b","Type":"ContainerDied","Data":"31d0cf5512a1cc5dd661ae6e504d43d09c1bc8e8da37b0a9967bc3e9d4d6db40"} Dec 09 09:53:05 crc kubenswrapper[4824]: I1209 09:53:05.867187 4824 generic.go:334] "Generic (PLEG): container finished" podID="844884cd-6813-4fa2-88e4-7a5994bbd3cb" containerID="8e40a8d1ab7f864d4b72e1ac8621c60082f7a380e8ce39e6d518520782dac70d" exitCode=0 Dec 09 09:53:05 crc kubenswrapper[4824]: I1209 09:53:05.867255 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"844884cd-6813-4fa2-88e4-7a5994bbd3cb","Type":"ContainerDied","Data":"8e40a8d1ab7f864d4b72e1ac8621c60082f7a380e8ce39e6d518520782dac70d"} Dec 09 09:53:05 crc kubenswrapper[4824]: I1209 09:53:05.867277 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"844884cd-6813-4fa2-88e4-7a5994bbd3cb","Type":"ContainerStarted","Data":"3c98e93226877ab72a03d11921a6055afb0a1956f7a43dfd88028ff5a6d79d1a"} Dec 09 09:53:05 crc kubenswrapper[4824]: I1209 09:53:05.870425 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-6c57467cd4-4q6wt" event={"ID":"0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3","Type":"ContainerStarted","Data":"3a50a62f2dc71b68d4700942b803396c3ba1b8b8963a9e2eb1c34924f3abd31f"} Dec 09 09:53:05 crc kubenswrapper[4824]: I1209 09:53:05.879377 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-684fc64449-59s5g" podStartSLOduration=4.879342964 podStartE2EDuration="4.879342964s" podCreationTimestamp="2025-12-09 09:53:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:53:05.878588701 +0000 UTC m=+342.213093388" watchObservedRunningTime="2025-12-09 09:53:05.879342964 +0000 UTC m=+342.213847631" Dec 09 09:53:06 crc kubenswrapper[4824]: I1209 09:53:06.015659 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/openshift-state-metrics-566fddb674-hjnrk" podStartSLOduration=4.764605047 podStartE2EDuration="10.015592386s" podCreationTimestamp="2025-12-09 09:52:56 +0000 UTC" firstStartedPulling="2025-12-09 09:52:59.32846301 +0000 UTC m=+335.662967677" lastFinishedPulling="2025-12-09 09:53:04.579450349 +0000 UTC m=+340.913955016" observedRunningTime="2025-12-09 09:53:06.014744709 +0000 UTC m=+342.349249386" watchObservedRunningTime="2025-12-09 09:53:06.015592386 +0000 UTC m=+342.350097053" Dec 09 09:53:11 crc kubenswrapper[4824]: I1209 09:53:11.678760 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:11 crc kubenswrapper[4824]: I1209 09:53:11.684744 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:11 crc kubenswrapper[4824]: I1209 09:53:11.699661 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:11 crc kubenswrapper[4824]: I1209 09:53:11.713587 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:53:11 crc kubenswrapper[4824]: I1209 09:53:11.810268 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-ldw7r"] Dec 09 09:53:13 crc kubenswrapper[4824]: I1209 09:53:13.988169 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"26c6a6ba-64bf-4473-816d-e9ee424bf97b","Type":"ContainerStarted","Data":"014152497d33bf91f6cb1659235aa1c2f255126f5c615d138ce234cfa17bf517"} Dec 09 09:53:13 crc kubenswrapper[4824]: I1209 09:53:13.989081 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"26c6a6ba-64bf-4473-816d-e9ee424bf97b","Type":"ContainerStarted","Data":"c3f952d734307d11bf6d32c23a3aa6e9652d424b9bdfcda9a9ea138ee50fa09d"} Dec 09 09:53:13 crc kubenswrapper[4824]: I1209 09:53:13.992877 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"844884cd-6813-4fa2-88e4-7a5994bbd3cb","Type":"ContainerStarted","Data":"b1b8eb4a7620dd8b672663cc1b3741fb7f00c059ea0d7c4fe107781e275df63e"} Dec 09 09:53:13 crc kubenswrapper[4824]: I1209 09:53:13.992920 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"844884cd-6813-4fa2-88e4-7a5994bbd3cb","Type":"ContainerStarted","Data":"6e17f3569f7bae03cb9216e9536d95ea5c483b72c92274695a26c529449060eb"} Dec 09 09:53:13 crc kubenswrapper[4824]: I1209 09:53:13.992933 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"844884cd-6813-4fa2-88e4-7a5994bbd3cb","Type":"ContainerStarted","Data":"158ac5fd42a072f74df2a57bf9f830e8b0bc5339d06db284e5e77f0149ab5951"} Dec 09 09:53:13 crc kubenswrapper[4824]: I1209 09:53:13.995295 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-6c57467cd4-4q6wt" event={"ID":"0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3","Type":"ContainerStarted","Data":"f1a817deaa3061d3646cc40088ce5badd857d0fbd7404821e4cdff7c61906a1e"} Dec 09 09:53:13 crc kubenswrapper[4824]: I1209 09:53:13.997016 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/monitoring-plugin-6c57467cd4-4q6wt" Dec 09 09:53:14 crc kubenswrapper[4824]: I1209 09:53:14.011911 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/monitoring-plugin-6c57467cd4-4q6wt" Dec 09 09:53:14 crc kubenswrapper[4824]: I1209 09:53:14.020844 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" event={"ID":"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6","Type":"ContainerStarted","Data":"28a8decd5a2989e3d840498e0b9eb031558170592ffd3244845bd4a8e96e90ec"} Dec 09 09:53:14 crc kubenswrapper[4824]: I1209 09:53:14.020910 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" event={"ID":"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6","Type":"ContainerStarted","Data":"e18f32f12f9cfdaaa3b3e3c05c5e05f8b3d3b9c986acfd426f3fe41f8074ca6b"} Dec 09 09:53:14 crc kubenswrapper[4824]: I1209 09:53:14.020923 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" event={"ID":"d030e6a9-d47a-47bf-9c24-ff0ca58f71e6","Type":"ContainerStarted","Data":"baa3b8cbd7223566a7664cd677b19649790fd21f751dce59951c7a6a9460eede"} Dec 09 09:53:14 crc kubenswrapper[4824]: I1209 09:53:14.021411 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:53:14 crc kubenswrapper[4824]: I1209 09:53:14.034135 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" event={"ID":"9aa0314b-a177-4589-bc32-f9caf6f8f724","Type":"ContainerStarted","Data":"c3e8d0870737ba4e5bbc0d028449d281500635ecaf1a9525b3d108cd03a45b6c"} Dec 09 09:53:14 crc kubenswrapper[4824]: I1209 09:53:14.050105 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" Dec 09 09:53:14 crc kubenswrapper[4824]: I1209 09:53:14.087732 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/monitoring-plugin-6c57467cd4-4q6wt" podStartSLOduration=3.622779465 podStartE2EDuration="11.08768182s" podCreationTimestamp="2025-12-09 09:53:03 +0000 UTC" firstStartedPulling="2025-12-09 09:53:05.298223535 +0000 UTC m=+341.632728202" lastFinishedPulling="2025-12-09 09:53:12.76312589 +0000 UTC m=+349.097630557" observedRunningTime="2025-12-09 09:53:14.082513318 +0000 UTC m=+350.417017985" watchObservedRunningTime="2025-12-09 09:53:14.08768182 +0000 UTC m=+350.422186487" Dec 09 09:53:14 crc kubenswrapper[4824]: I1209 09:53:14.142146 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" podStartSLOduration=5.7634632759999995 podStartE2EDuration="13.142111671s" podCreationTimestamp="2025-12-09 09:53:01 +0000 UTC" firstStartedPulling="2025-12-09 09:53:05.388089598 +0000 UTC m=+341.722594265" lastFinishedPulling="2025-12-09 09:53:12.766737993 +0000 UTC m=+349.101242660" observedRunningTime="2025-12-09 09:53:14.131426465 +0000 UTC m=+350.465931152" watchObservedRunningTime="2025-12-09 09:53:14.142111671 +0000 UTC m=+350.476616338" Dec 09 09:53:14 crc kubenswrapper[4824]: I1209 09:53:14.196234 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" podStartSLOduration=3.616036351 podStartE2EDuration="16.196205071s" podCreationTimestamp="2025-12-09 09:52:58 +0000 UTC" firstStartedPulling="2025-12-09 09:53:00.18295355 +0000 UTC m=+336.517458217" lastFinishedPulling="2025-12-09 09:53:12.76312227 +0000 UTC m=+349.097626937" observedRunningTime="2025-12-09 09:53:14.193839526 +0000 UTC m=+350.528344223" watchObservedRunningTime="2025-12-09 09:53:14.196205071 +0000 UTC m=+350.530709738" Dec 09 09:53:15 crc kubenswrapper[4824]: I1209 09:53:15.079024 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"26c6a6ba-64bf-4473-816d-e9ee424bf97b","Type":"ContainerStarted","Data":"63a04e956ed65d9e6e8f32ff9ff8217f5bf1ede6660ab785eff0c64a397a7b65"} Dec 09 09:53:15 crc kubenswrapper[4824]: I1209 09:53:15.079097 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"26c6a6ba-64bf-4473-816d-e9ee424bf97b","Type":"ContainerStarted","Data":"e186dd8abc81d74f5609256d95df86bd276d27480a78ffe43086aea588435349"} Dec 09 09:53:15 crc kubenswrapper[4824]: I1209 09:53:15.079114 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"26c6a6ba-64bf-4473-816d-e9ee424bf97b","Type":"ContainerStarted","Data":"b01cffdfb5130a1c0f932dcf80733174cfdc6fb9927f0b06565af82532681abe"} Dec 09 09:53:15 crc kubenswrapper[4824]: I1209 09:53:15.079128 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"26c6a6ba-64bf-4473-816d-e9ee424bf97b","Type":"ContainerStarted","Data":"a58af6ce530074fc8e0aed1c6674be131ad1bd2e10a4aad305d1546f885fc423"} Dec 09 09:53:15 crc kubenswrapper[4824]: I1209 09:53:15.099084 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"844884cd-6813-4fa2-88e4-7a5994bbd3cb","Type":"ContainerStarted","Data":"e8397b0f3efd7f511b875e0d43dff078f8f66b046e75782109659178fe64eefb"} Dec 09 09:53:15 crc kubenswrapper[4824]: I1209 09:53:15.099146 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"844884cd-6813-4fa2-88e4-7a5994bbd3cb","Type":"ContainerStarted","Data":"b280cc1f6369d03a2f1feda1a9ba25eb7fd7751643d8225bd1b68fa0253f0be0"} Dec 09 09:53:15 crc kubenswrapper[4824]: I1209 09:53:15.128512 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/alertmanager-main-0" podStartSLOduration=4.811553556 podStartE2EDuration="18.128487266s" podCreationTimestamp="2025-12-09 09:52:57 +0000 UTC" firstStartedPulling="2025-12-09 09:52:59.492180235 +0000 UTC m=+335.826684902" lastFinishedPulling="2025-12-09 09:53:12.809113945 +0000 UTC m=+349.143618612" observedRunningTime="2025-12-09 09:53:15.122538618 +0000 UTC m=+351.457043295" watchObservedRunningTime="2025-12-09 09:53:15.128487266 +0000 UTC m=+351.462991923" Dec 09 09:53:15 crc kubenswrapper[4824]: I1209 09:53:15.185064 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-k8s-0" podStartSLOduration=6.223158264 podStartE2EDuration="13.185038952s" podCreationTimestamp="2025-12-09 09:53:02 +0000 UTC" firstStartedPulling="2025-12-09 09:53:05.868735541 +0000 UTC m=+342.203256749" lastFinishedPulling="2025-12-09 09:53:12.83063277 +0000 UTC m=+349.165137437" observedRunningTime="2025-12-09 09:53:15.179307712 +0000 UTC m=+351.513812389" watchObservedRunningTime="2025-12-09 09:53:15.185038952 +0000 UTC m=+351.519543619" Dec 09 09:53:16 crc kubenswrapper[4824]: I1209 09:53:16.110528 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"844884cd-6813-4fa2-88e4-7a5994bbd3cb","Type":"ContainerStarted","Data":"e6f847b9bc3e2d4c629b179b742b615f076e2a4e59726ce7df9a77bbbbcbbc02"} Dec 09 09:53:18 crc kubenswrapper[4824]: I1209 09:53:18.905158 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:53:20 crc kubenswrapper[4824]: I1209 09:53:20.392939 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" Dec 09 09:53:20 crc kubenswrapper[4824]: I1209 09:53:20.468296 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qkkdx"] Dec 09 09:53:22 crc kubenswrapper[4824]: I1209 09:53:22.267612 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:22 crc kubenswrapper[4824]: I1209 09:53:22.268284 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:32 crc kubenswrapper[4824]: I1209 09:53:32.861024 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 09:53:32 crc kubenswrapper[4824]: I1209 09:53:32.861899 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 09:53:36 crc kubenswrapper[4824]: I1209 09:53:36.882239 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-ldw7r" podUID="0c5e7133-21dc-46e0-b334-9d36c0427aa3" containerName="console" containerID="cri-o://5156e771cd0edbb835f34d16e3d11ec265dc92c76c7dce49b404389fcf18d2b7" gracePeriod=15 Dec 09 09:53:37 crc kubenswrapper[4824]: I1209 09:53:37.824943 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-ldw7r_0c5e7133-21dc-46e0-b334-9d36c0427aa3/console/0.log" Dec 09 09:53:37 crc kubenswrapper[4824]: I1209 09:53:37.825332 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:53:37 crc kubenswrapper[4824]: I1209 09:53:37.879281 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0c5e7133-21dc-46e0-b334-9d36c0427aa3-oauth-serving-cert\") pod \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " Dec 09 09:53:37 crc kubenswrapper[4824]: I1209 09:53:37.879348 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0c5e7133-21dc-46e0-b334-9d36c0427aa3-console-serving-cert\") pod \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " Dec 09 09:53:37 crc kubenswrapper[4824]: I1209 09:53:37.879393 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0c5e7133-21dc-46e0-b334-9d36c0427aa3-service-ca\") pod \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " Dec 09 09:53:37 crc kubenswrapper[4824]: I1209 09:53:37.879483 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdcvb\" (UniqueName: \"kubernetes.io/projected/0c5e7133-21dc-46e0-b334-9d36c0427aa3-kube-api-access-zdcvb\") pod \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " Dec 09 09:53:37 crc kubenswrapper[4824]: I1209 09:53:37.879534 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0c5e7133-21dc-46e0-b334-9d36c0427aa3-console-oauth-config\") pod \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " Dec 09 09:53:37 crc kubenswrapper[4824]: I1209 09:53:37.879574 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0c5e7133-21dc-46e0-b334-9d36c0427aa3-trusted-ca-bundle\") pod \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " Dec 09 09:53:37 crc kubenswrapper[4824]: I1209 09:53:37.879602 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0c5e7133-21dc-46e0-b334-9d36c0427aa3-console-config\") pod \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\" (UID: \"0c5e7133-21dc-46e0-b334-9d36c0427aa3\") " Dec 09 09:53:37 crc kubenswrapper[4824]: I1209 09:53:37.880338 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c5e7133-21dc-46e0-b334-9d36c0427aa3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0c5e7133-21dc-46e0-b334-9d36c0427aa3" (UID: "0c5e7133-21dc-46e0-b334-9d36c0427aa3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:53:37 crc kubenswrapper[4824]: I1209 09:53:37.880587 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c5e7133-21dc-46e0-b334-9d36c0427aa3-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "0c5e7133-21dc-46e0-b334-9d36c0427aa3" (UID: "0c5e7133-21dc-46e0-b334-9d36c0427aa3"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:53:37 crc kubenswrapper[4824]: I1209 09:53:37.880632 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c5e7133-21dc-46e0-b334-9d36c0427aa3-console-config" (OuterVolumeSpecName: "console-config") pod "0c5e7133-21dc-46e0-b334-9d36c0427aa3" (UID: "0c5e7133-21dc-46e0-b334-9d36c0427aa3"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:53:37 crc kubenswrapper[4824]: I1209 09:53:37.880578 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c5e7133-21dc-46e0-b334-9d36c0427aa3-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "0c5e7133-21dc-46e0-b334-9d36c0427aa3" (UID: "0c5e7133-21dc-46e0-b334-9d36c0427aa3"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:53:37 crc kubenswrapper[4824]: I1209 09:53:37.886730 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c5e7133-21dc-46e0-b334-9d36c0427aa3-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "0c5e7133-21dc-46e0-b334-9d36c0427aa3" (UID: "0c5e7133-21dc-46e0-b334-9d36c0427aa3"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:53:37 crc kubenswrapper[4824]: I1209 09:53:37.886848 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c5e7133-21dc-46e0-b334-9d36c0427aa3-kube-api-access-zdcvb" (OuterVolumeSpecName: "kube-api-access-zdcvb") pod "0c5e7133-21dc-46e0-b334-9d36c0427aa3" (UID: "0c5e7133-21dc-46e0-b334-9d36c0427aa3"). InnerVolumeSpecName "kube-api-access-zdcvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:53:37 crc kubenswrapper[4824]: I1209 09:53:37.887050 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c5e7133-21dc-46e0-b334-9d36c0427aa3-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "0c5e7133-21dc-46e0-b334-9d36c0427aa3" (UID: "0c5e7133-21dc-46e0-b334-9d36c0427aa3"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:53:37 crc kubenswrapper[4824]: I1209 09:53:37.982143 4824 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0c5e7133-21dc-46e0-b334-9d36c0427aa3-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:53:37 crc kubenswrapper[4824]: I1209 09:53:37.982293 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0c5e7133-21dc-46e0-b334-9d36c0427aa3-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 09:53:37 crc kubenswrapper[4824]: I1209 09:53:37.982314 4824 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0c5e7133-21dc-46e0-b334-9d36c0427aa3-console-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:53:37 crc kubenswrapper[4824]: I1209 09:53:37.982326 4824 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0c5e7133-21dc-46e0-b334-9d36c0427aa3-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:53:37 crc kubenswrapper[4824]: I1209 09:53:37.982358 4824 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0c5e7133-21dc-46e0-b334-9d36c0427aa3-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:53:37 crc kubenswrapper[4824]: I1209 09:53:37.982372 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0c5e7133-21dc-46e0-b334-9d36c0427aa3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 09:53:37 crc kubenswrapper[4824]: I1209 09:53:37.982387 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdcvb\" (UniqueName: \"kubernetes.io/projected/0c5e7133-21dc-46e0-b334-9d36c0427aa3-kube-api-access-zdcvb\") on node \"crc\" DevicePath \"\"" Dec 09 09:53:38 crc kubenswrapper[4824]: I1209 09:53:38.270043 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-ldw7r_0c5e7133-21dc-46e0-b334-9d36c0427aa3/console/0.log" Dec 09 09:53:38 crc kubenswrapper[4824]: I1209 09:53:38.270506 4824 generic.go:334] "Generic (PLEG): container finished" podID="0c5e7133-21dc-46e0-b334-9d36c0427aa3" containerID="5156e771cd0edbb835f34d16e3d11ec265dc92c76c7dce49b404389fcf18d2b7" exitCode=2 Dec 09 09:53:38 crc kubenswrapper[4824]: I1209 09:53:38.270584 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ldw7r" event={"ID":"0c5e7133-21dc-46e0-b334-9d36c0427aa3","Type":"ContainerDied","Data":"5156e771cd0edbb835f34d16e3d11ec265dc92c76c7dce49b404389fcf18d2b7"} Dec 09 09:53:38 crc kubenswrapper[4824]: I1209 09:53:38.270615 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ldw7r" Dec 09 09:53:38 crc kubenswrapper[4824]: I1209 09:53:38.270657 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ldw7r" event={"ID":"0c5e7133-21dc-46e0-b334-9d36c0427aa3","Type":"ContainerDied","Data":"800371374fe33dfa445ea3c80d7fb2b4d04e84086246d829bf952bc1cd47e079"} Dec 09 09:53:38 crc kubenswrapper[4824]: I1209 09:53:38.270694 4824 scope.go:117] "RemoveContainer" containerID="5156e771cd0edbb835f34d16e3d11ec265dc92c76c7dce49b404389fcf18d2b7" Dec 09 09:53:38 crc kubenswrapper[4824]: I1209 09:53:38.293940 4824 scope.go:117] "RemoveContainer" containerID="5156e771cd0edbb835f34d16e3d11ec265dc92c76c7dce49b404389fcf18d2b7" Dec 09 09:53:38 crc kubenswrapper[4824]: E1209 09:53:38.294555 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5156e771cd0edbb835f34d16e3d11ec265dc92c76c7dce49b404389fcf18d2b7\": container with ID starting with 5156e771cd0edbb835f34d16e3d11ec265dc92c76c7dce49b404389fcf18d2b7 not found: ID does not exist" containerID="5156e771cd0edbb835f34d16e3d11ec265dc92c76c7dce49b404389fcf18d2b7" Dec 09 09:53:38 crc kubenswrapper[4824]: I1209 09:53:38.294599 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5156e771cd0edbb835f34d16e3d11ec265dc92c76c7dce49b404389fcf18d2b7"} err="failed to get container status \"5156e771cd0edbb835f34d16e3d11ec265dc92c76c7dce49b404389fcf18d2b7\": rpc error: code = NotFound desc = could not find container \"5156e771cd0edbb835f34d16e3d11ec265dc92c76c7dce49b404389fcf18d2b7\": container with ID starting with 5156e771cd0edbb835f34d16e3d11ec265dc92c76c7dce49b404389fcf18d2b7 not found: ID does not exist" Dec 09 09:53:38 crc kubenswrapper[4824]: I1209 09:53:38.295525 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-ldw7r"] Dec 09 09:53:38 crc kubenswrapper[4824]: I1209 09:53:38.302939 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-ldw7r"] Dec 09 09:53:39 crc kubenswrapper[4824]: I1209 09:53:39.918709 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c5e7133-21dc-46e0-b334-9d36c0427aa3" path="/var/lib/kubelet/pods/0c5e7133-21dc-46e0-b334-9d36c0427aa3/volumes" Dec 09 09:53:42 crc kubenswrapper[4824]: I1209 09:53:42.275290 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:42 crc kubenswrapper[4824]: I1209 09:53:42.280715 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:53:45 crc kubenswrapper[4824]: I1209 09:53:45.518602 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" podUID="6d910d95-4961-4d28-9599-087153ce7d2f" containerName="registry" containerID="cri-o://d197cdd6496cac6eb5e8db0458314c964b45b0ebe8bf001ebc5b60eee8a75e47" gracePeriod=30 Dec 09 09:53:45 crc kubenswrapper[4824]: I1209 09:53:45.572506 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6ffd5d66b-c8452"] Dec 09 09:53:45 crc kubenswrapper[4824]: I1209 09:53:45.573294 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" podUID="bcaf42e7-e583-4767-ac14-a48f7160823a" containerName="controller-manager" containerID="cri-o://d237271ae948f8b60cb86d6c32f4ae74e4e577c581f4511931de72d6973f936e" gracePeriod=30 Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.037132 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.112982 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"6d910d95-4961-4d28-9599-087153ce7d2f\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.113034 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6d910d95-4961-4d28-9599-087153ce7d2f-registry-certificates\") pod \"6d910d95-4961-4d28-9599-087153ce7d2f\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.113087 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6d910d95-4961-4d28-9599-087153ce7d2f-trusted-ca\") pod \"6d910d95-4961-4d28-9599-087153ce7d2f\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.113140 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6d910d95-4961-4d28-9599-087153ce7d2f-bound-sa-token\") pod \"6d910d95-4961-4d28-9599-087153ce7d2f\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.113174 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6d910d95-4961-4d28-9599-087153ce7d2f-registry-tls\") pod \"6d910d95-4961-4d28-9599-087153ce7d2f\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.115084 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d910d95-4961-4d28-9599-087153ce7d2f-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "6d910d95-4961-4d28-9599-087153ce7d2f" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.116136 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d910d95-4961-4d28-9599-087153ce7d2f-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "6d910d95-4961-4d28-9599-087153ce7d2f" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.122883 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d910d95-4961-4d28-9599-087153ce7d2f-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "6d910d95-4961-4d28-9599-087153ce7d2f" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.124436 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d910d95-4961-4d28-9599-087153ce7d2f-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "6d910d95-4961-4d28-9599-087153ce7d2f" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.124850 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "6d910d95-4961-4d28-9599-087153ce7d2f" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.154231 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.226310 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kthx2\" (UniqueName: \"kubernetes.io/projected/6d910d95-4961-4d28-9599-087153ce7d2f-kube-api-access-kthx2\") pod \"6d910d95-4961-4d28-9599-087153ce7d2f\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.226409 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6d910d95-4961-4d28-9599-087153ce7d2f-ca-trust-extracted\") pod \"6d910d95-4961-4d28-9599-087153ce7d2f\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.226447 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6d910d95-4961-4d28-9599-087153ce7d2f-installation-pull-secrets\") pod \"6d910d95-4961-4d28-9599-087153ce7d2f\" (UID: \"6d910d95-4961-4d28-9599-087153ce7d2f\") " Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.226917 4824 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6d910d95-4961-4d28-9599-087153ce7d2f-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.226949 4824 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6d910d95-4961-4d28-9599-087153ce7d2f-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.226969 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6d910d95-4961-4d28-9599-087153ce7d2f-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.226986 4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6d910d95-4961-4d28-9599-087153ce7d2f-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.233114 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d910d95-4961-4d28-9599-087153ce7d2f-kube-api-access-kthx2" (OuterVolumeSpecName: "kube-api-access-kthx2") pod "6d910d95-4961-4d28-9599-087153ce7d2f" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f"). InnerVolumeSpecName "kube-api-access-kthx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.234622 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d910d95-4961-4d28-9599-087153ce7d2f-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "6d910d95-4961-4d28-9599-087153ce7d2f" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.259609 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d910d95-4961-4d28-9599-087153ce7d2f-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "6d910d95-4961-4d28-9599-087153ce7d2f" (UID: "6d910d95-4961-4d28-9599-087153ce7d2f"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.328178 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bcaf42e7-e583-4767-ac14-a48f7160823a-proxy-ca-bundles\") pod \"bcaf42e7-e583-4767-ac14-a48f7160823a\" (UID: \"bcaf42e7-e583-4767-ac14-a48f7160823a\") " Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.329337 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcaf42e7-e583-4767-ac14-a48f7160823a-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "bcaf42e7-e583-4767-ac14-a48f7160823a" (UID: "bcaf42e7-e583-4767-ac14-a48f7160823a"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.330238 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcaf42e7-e583-4767-ac14-a48f7160823a-client-ca" (OuterVolumeSpecName: "client-ca") pod "bcaf42e7-e583-4767-ac14-a48f7160823a" (UID: "bcaf42e7-e583-4767-ac14-a48f7160823a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.330295 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bcaf42e7-e583-4767-ac14-a48f7160823a-client-ca\") pod \"bcaf42e7-e583-4767-ac14-a48f7160823a\" (UID: \"bcaf42e7-e583-4767-ac14-a48f7160823a\") " Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.330461 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cd469\" (UniqueName: \"kubernetes.io/projected/bcaf42e7-e583-4767-ac14-a48f7160823a-kube-api-access-cd469\") pod \"bcaf42e7-e583-4767-ac14-a48f7160823a\" (UID: \"bcaf42e7-e583-4767-ac14-a48f7160823a\") " Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.331124 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcaf42e7-e583-4767-ac14-a48f7160823a-config\") pod \"bcaf42e7-e583-4767-ac14-a48f7160823a\" (UID: \"bcaf42e7-e583-4767-ac14-a48f7160823a\") " Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.331177 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bcaf42e7-e583-4767-ac14-a48f7160823a-serving-cert\") pod \"bcaf42e7-e583-4767-ac14-a48f7160823a\" (UID: \"bcaf42e7-e583-4767-ac14-a48f7160823a\") " Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.331628 4824 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6d910d95-4961-4d28-9599-087153ce7d2f-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.331643 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bcaf42e7-e583-4767-ac14-a48f7160823a-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.331654 4824 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6d910d95-4961-4d28-9599-087153ce7d2f-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.331667 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bcaf42e7-e583-4767-ac14-a48f7160823a-client-ca\") on node \"crc\" DevicePath \"\"" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.331679 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kthx2\" (UniqueName: \"kubernetes.io/projected/6d910d95-4961-4d28-9599-087153ce7d2f-kube-api-access-kthx2\") on node \"crc\" DevicePath \"\"" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.332277 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcaf42e7-e583-4767-ac14-a48f7160823a-config" (OuterVolumeSpecName: "config") pod "bcaf42e7-e583-4767-ac14-a48f7160823a" (UID: "bcaf42e7-e583-4767-ac14-a48f7160823a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.334388 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcaf42e7-e583-4767-ac14-a48f7160823a-kube-api-access-cd469" (OuterVolumeSpecName: "kube-api-access-cd469") pod "bcaf42e7-e583-4767-ac14-a48f7160823a" (UID: "bcaf42e7-e583-4767-ac14-a48f7160823a"). InnerVolumeSpecName "kube-api-access-cd469". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.336822 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcaf42e7-e583-4767-ac14-a48f7160823a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bcaf42e7-e583-4767-ac14-a48f7160823a" (UID: "bcaf42e7-e583-4767-ac14-a48f7160823a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.341695 4824 generic.go:334] "Generic (PLEG): container finished" podID="6d910d95-4961-4d28-9599-087153ce7d2f" containerID="d197cdd6496cac6eb5e8db0458314c964b45b0ebe8bf001ebc5b60eee8a75e47" exitCode=0 Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.341808 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" event={"ID":"6d910d95-4961-4d28-9599-087153ce7d2f","Type":"ContainerDied","Data":"d197cdd6496cac6eb5e8db0458314c964b45b0ebe8bf001ebc5b60eee8a75e47"} Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.342337 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" event={"ID":"6d910d95-4961-4d28-9599-087153ce7d2f","Type":"ContainerDied","Data":"1b5175598a7d3f9dc435aec37623ff6fed9118de27a47b9b4509953362a3348a"} Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.341825 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qkkdx" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.342385 4824 scope.go:117] "RemoveContainer" containerID="d197cdd6496cac6eb5e8db0458314c964b45b0ebe8bf001ebc5b60eee8a75e47" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.344773 4824 generic.go:334] "Generic (PLEG): container finished" podID="bcaf42e7-e583-4767-ac14-a48f7160823a" containerID="d237271ae948f8b60cb86d6c32f4ae74e4e577c581f4511931de72d6973f936e" exitCode=0 Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.344859 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" event={"ID":"bcaf42e7-e583-4767-ac14-a48f7160823a","Type":"ContainerDied","Data":"d237271ae948f8b60cb86d6c32f4ae74e4e577c581f4511931de72d6973f936e"} Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.345024 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.346058 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6ffd5d66b-c8452" event={"ID":"bcaf42e7-e583-4767-ac14-a48f7160823a","Type":"ContainerDied","Data":"0db146f95ea4b3abc52c1489781f10403aa73f710724aa5d09ca689f7f081156"} Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.377881 4824 scope.go:117] "RemoveContainer" containerID="d197cdd6496cac6eb5e8db0458314c964b45b0ebe8bf001ebc5b60eee8a75e47" Dec 09 09:53:47 crc kubenswrapper[4824]: E1209 09:53:47.378686 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d197cdd6496cac6eb5e8db0458314c964b45b0ebe8bf001ebc5b60eee8a75e47\": container with ID starting with d197cdd6496cac6eb5e8db0458314c964b45b0ebe8bf001ebc5b60eee8a75e47 not found: ID does not exist" containerID="d197cdd6496cac6eb5e8db0458314c964b45b0ebe8bf001ebc5b60eee8a75e47" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.378760 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d197cdd6496cac6eb5e8db0458314c964b45b0ebe8bf001ebc5b60eee8a75e47"} err="failed to get container status \"d197cdd6496cac6eb5e8db0458314c964b45b0ebe8bf001ebc5b60eee8a75e47\": rpc error: code = NotFound desc = could not find container \"d197cdd6496cac6eb5e8db0458314c964b45b0ebe8bf001ebc5b60eee8a75e47\": container with ID starting with d197cdd6496cac6eb5e8db0458314c964b45b0ebe8bf001ebc5b60eee8a75e47 not found: ID does not exist" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.378835 4824 scope.go:117] "RemoveContainer" containerID="d237271ae948f8b60cb86d6c32f4ae74e4e577c581f4511931de72d6973f936e" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.386813 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qkkdx"] Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.399099 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qkkdx"] Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.408124 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6ffd5d66b-c8452"] Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.412200 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6ffd5d66b-c8452"] Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.412751 4824 scope.go:117] "RemoveContainer" containerID="d237271ae948f8b60cb86d6c32f4ae74e4e577c581f4511931de72d6973f936e" Dec 09 09:53:47 crc kubenswrapper[4824]: E1209 09:53:47.413472 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d237271ae948f8b60cb86d6c32f4ae74e4e577c581f4511931de72d6973f936e\": container with ID starting with d237271ae948f8b60cb86d6c32f4ae74e4e577c581f4511931de72d6973f936e not found: ID does not exist" containerID="d237271ae948f8b60cb86d6c32f4ae74e4e577c581f4511931de72d6973f936e" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.413562 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d237271ae948f8b60cb86d6c32f4ae74e4e577c581f4511931de72d6973f936e"} err="failed to get container status \"d237271ae948f8b60cb86d6c32f4ae74e4e577c581f4511931de72d6973f936e\": rpc error: code = NotFound desc = could not find container \"d237271ae948f8b60cb86d6c32f4ae74e4e577c581f4511931de72d6973f936e\": container with ID starting with d237271ae948f8b60cb86d6c32f4ae74e4e577c581f4511931de72d6973f936e not found: ID does not exist" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.432954 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cd469\" (UniqueName: \"kubernetes.io/projected/bcaf42e7-e583-4767-ac14-a48f7160823a-kube-api-access-cd469\") on node \"crc\" DevicePath \"\"" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.433298 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcaf42e7-e583-4767-ac14-a48f7160823a-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.433394 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bcaf42e7-e583-4767-ac14-a48f7160823a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.919579 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d910d95-4961-4d28-9599-087153ce7d2f" path="/var/lib/kubelet/pods/6d910d95-4961-4d28-9599-087153ce7d2f/volumes" Dec 09 09:53:47 crc kubenswrapper[4824]: I1209 09:53:47.920216 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcaf42e7-e583-4767-ac14-a48f7160823a" path="/var/lib/kubelet/pods/bcaf42e7-e583-4767-ac14-a48f7160823a/volumes" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.788285 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-69949c7bd4-pmmk2"] Dec 09 09:53:49 crc kubenswrapper[4824]: E1209 09:53:49.789072 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c5e7133-21dc-46e0-b334-9d36c0427aa3" containerName="console" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.789089 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c5e7133-21dc-46e0-b334-9d36c0427aa3" containerName="console" Dec 09 09:53:49 crc kubenswrapper[4824]: E1209 09:53:49.789123 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d910d95-4961-4d28-9599-087153ce7d2f" containerName="registry" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.789130 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d910d95-4961-4d28-9599-087153ce7d2f" containerName="registry" Dec 09 09:53:49 crc kubenswrapper[4824]: E1209 09:53:49.789143 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcaf42e7-e583-4767-ac14-a48f7160823a" containerName="controller-manager" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.789152 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcaf42e7-e583-4767-ac14-a48f7160823a" containerName="controller-manager" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.789268 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcaf42e7-e583-4767-ac14-a48f7160823a" containerName="controller-manager" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.789284 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c5e7133-21dc-46e0-b334-9d36c0427aa3" containerName="console" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.789298 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d910d95-4961-4d28-9599-087153ce7d2f" containerName="registry" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.789991 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.793024 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.793078 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.793251 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.793024 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.802423 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-69949c7bd4-pmmk2"] Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.806117 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.806352 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.806485 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.876145 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc-client-ca\") pod \"controller-manager-69949c7bd4-pmmk2\" (UID: \"ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc\") " pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.876215 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc-serving-cert\") pod \"controller-manager-69949c7bd4-pmmk2\" (UID: \"ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc\") " pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.876377 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc-config\") pod \"controller-manager-69949c7bd4-pmmk2\" (UID: \"ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc\") " pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.876465 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc-proxy-ca-bundles\") pod \"controller-manager-69949c7bd4-pmmk2\" (UID: \"ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc\") " pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.876656 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6snbk\" (UniqueName: \"kubernetes.io/projected/ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc-kube-api-access-6snbk\") pod \"controller-manager-69949c7bd4-pmmk2\" (UID: \"ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc\") " pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.979127 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc-client-ca\") pod \"controller-manager-69949c7bd4-pmmk2\" (UID: \"ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc\") " pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.979224 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc-serving-cert\") pod \"controller-manager-69949c7bd4-pmmk2\" (UID: \"ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc\") " pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.979280 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc-config\") pod \"controller-manager-69949c7bd4-pmmk2\" (UID: \"ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc\") " pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.979331 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc-proxy-ca-bundles\") pod \"controller-manager-69949c7bd4-pmmk2\" (UID: \"ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc\") " pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.979470 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6snbk\" (UniqueName: \"kubernetes.io/projected/ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc-kube-api-access-6snbk\") pod \"controller-manager-69949c7bd4-pmmk2\" (UID: \"ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc\") " pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.980367 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc-client-ca\") pod \"controller-manager-69949c7bd4-pmmk2\" (UID: \"ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc\") " pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.980866 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc-config\") pod \"controller-manager-69949c7bd4-pmmk2\" (UID: \"ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc\") " pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.981490 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc-proxy-ca-bundles\") pod \"controller-manager-69949c7bd4-pmmk2\" (UID: \"ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc\") " pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.988403 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc-serving-cert\") pod \"controller-manager-69949c7bd4-pmmk2\" (UID: \"ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc\") " pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" Dec 09 09:53:49 crc kubenswrapper[4824]: I1209 09:53:49.999392 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6snbk\" (UniqueName: \"kubernetes.io/projected/ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc-kube-api-access-6snbk\") pod \"controller-manager-69949c7bd4-pmmk2\" (UID: \"ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc\") " pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" Dec 09 09:53:50 crc kubenswrapper[4824]: I1209 09:53:50.131850 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" Dec 09 09:53:50 crc kubenswrapper[4824]: I1209 09:53:50.363957 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-69949c7bd4-pmmk2"] Dec 09 09:53:51 crc kubenswrapper[4824]: I1209 09:53:51.380372 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" event={"ID":"ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc","Type":"ContainerStarted","Data":"8ef0f4d72f39e4c91f5d814ad0f5248f69fc7641a71cf06a32de5125e7d45dc5"} Dec 09 09:53:51 crc kubenswrapper[4824]: I1209 09:53:51.382307 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" event={"ID":"ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc","Type":"ContainerStarted","Data":"c6d5a291fa4a6597f17a400c146ba588eb0b96c016d7f2775ba989612db874c2"} Dec 09 09:53:51 crc kubenswrapper[4824]: I1209 09:53:51.382455 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" Dec 09 09:53:51 crc kubenswrapper[4824]: I1209 09:53:51.387858 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" Dec 09 09:53:51 crc kubenswrapper[4824]: I1209 09:53:51.402389 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" podStartSLOduration=6.402362666 podStartE2EDuration="6.402362666s" podCreationTimestamp="2025-12-09 09:53:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:53:51.401614482 +0000 UTC m=+387.736119159" watchObservedRunningTime="2025-12-09 09:53:51.402362666 +0000 UTC m=+387.736867333" Dec 09 09:54:02 crc kubenswrapper[4824]: I1209 09:54:02.861428 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 09:54:02 crc kubenswrapper[4824]: I1209 09:54:02.862304 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 09:54:03 crc kubenswrapper[4824]: I1209 09:54:03.946625 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:54:04 crc kubenswrapper[4824]: I1209 09:54:04.202123 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:54:04 crc kubenswrapper[4824]: I1209 09:54:04.510267 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Dec 09 09:54:20 crc kubenswrapper[4824]: I1209 09:54:20.886857 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/metrics-server-65846dd9c4-g8tlp"] Dec 09 09:54:20 crc kubenswrapper[4824]: I1209 09:54:20.889114 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:54:20 crc kubenswrapper[4824]: I1209 09:54:20.892041 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-e9oa5gmcvsrom" Dec 09 09:54:20 crc kubenswrapper[4824]: I1209 09:54:20.894935 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-monitoring/metrics-server-6448d488d4-xnf9s"] Dec 09 09:54:20 crc kubenswrapper[4824]: I1209 09:54:20.895290 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" podUID="9aa0314b-a177-4589-bc32-f9caf6f8f724" containerName="metrics-server" containerID="cri-o://c3e8d0870737ba4e5bbc0d028449d281500635ecaf1a9525b3d108cd03a45b6c" gracePeriod=170 Dec 09 09:54:20 crc kubenswrapper[4824]: I1209 09:54:20.919837 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-65846dd9c4-g8tlp"] Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.085123 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e153e30f-4f5e-4a38-8efb-6452096e25d7-client-ca-bundle\") pod \"metrics-server-65846dd9c4-g8tlp\" (UID: \"e153e30f-4f5e-4a38-8efb-6452096e25d7\") " pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.085296 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/e153e30f-4f5e-4a38-8efb-6452096e25d7-secret-metrics-client-certs\") pod \"metrics-server-65846dd9c4-g8tlp\" (UID: \"e153e30f-4f5e-4a38-8efb-6452096e25d7\") " pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.085359 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/e153e30f-4f5e-4a38-8efb-6452096e25d7-secret-metrics-server-tls\") pod \"metrics-server-65846dd9c4-g8tlp\" (UID: \"e153e30f-4f5e-4a38-8efb-6452096e25d7\") " pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.085384 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e153e30f-4f5e-4a38-8efb-6452096e25d7-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-65846dd9c4-g8tlp\" (UID: \"e153e30f-4f5e-4a38-8efb-6452096e25d7\") " pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.085410 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xsqz\" (UniqueName: \"kubernetes.io/projected/e153e30f-4f5e-4a38-8efb-6452096e25d7-kube-api-access-7xsqz\") pod \"metrics-server-65846dd9c4-g8tlp\" (UID: \"e153e30f-4f5e-4a38-8efb-6452096e25d7\") " pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.086633 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/e153e30f-4f5e-4a38-8efb-6452096e25d7-audit-log\") pod \"metrics-server-65846dd9c4-g8tlp\" (UID: \"e153e30f-4f5e-4a38-8efb-6452096e25d7\") " pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.086740 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/e153e30f-4f5e-4a38-8efb-6452096e25d7-metrics-server-audit-profiles\") pod \"metrics-server-65846dd9c4-g8tlp\" (UID: \"e153e30f-4f5e-4a38-8efb-6452096e25d7\") " pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.189275 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e153e30f-4f5e-4a38-8efb-6452096e25d7-client-ca-bundle\") pod \"metrics-server-65846dd9c4-g8tlp\" (UID: \"e153e30f-4f5e-4a38-8efb-6452096e25d7\") " pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.189379 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/e153e30f-4f5e-4a38-8efb-6452096e25d7-secret-metrics-client-certs\") pod \"metrics-server-65846dd9c4-g8tlp\" (UID: \"e153e30f-4f5e-4a38-8efb-6452096e25d7\") " pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.189446 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/e153e30f-4f5e-4a38-8efb-6452096e25d7-secret-metrics-server-tls\") pod \"metrics-server-65846dd9c4-g8tlp\" (UID: \"e153e30f-4f5e-4a38-8efb-6452096e25d7\") " pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.189469 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e153e30f-4f5e-4a38-8efb-6452096e25d7-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-65846dd9c4-g8tlp\" (UID: \"e153e30f-4f5e-4a38-8efb-6452096e25d7\") " pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.189489 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xsqz\" (UniqueName: \"kubernetes.io/projected/e153e30f-4f5e-4a38-8efb-6452096e25d7-kube-api-access-7xsqz\") pod \"metrics-server-65846dd9c4-g8tlp\" (UID: \"e153e30f-4f5e-4a38-8efb-6452096e25d7\") " pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.189516 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/e153e30f-4f5e-4a38-8efb-6452096e25d7-audit-log\") pod \"metrics-server-65846dd9c4-g8tlp\" (UID: \"e153e30f-4f5e-4a38-8efb-6452096e25d7\") " pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.189548 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/e153e30f-4f5e-4a38-8efb-6452096e25d7-metrics-server-audit-profiles\") pod \"metrics-server-65846dd9c4-g8tlp\" (UID: \"e153e30f-4f5e-4a38-8efb-6452096e25d7\") " pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.190423 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/e153e30f-4f5e-4a38-8efb-6452096e25d7-audit-log\") pod \"metrics-server-65846dd9c4-g8tlp\" (UID: \"e153e30f-4f5e-4a38-8efb-6452096e25d7\") " pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.191127 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/e153e30f-4f5e-4a38-8efb-6452096e25d7-metrics-server-audit-profiles\") pod \"metrics-server-65846dd9c4-g8tlp\" (UID: \"e153e30f-4f5e-4a38-8efb-6452096e25d7\") " pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.191659 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e153e30f-4f5e-4a38-8efb-6452096e25d7-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-65846dd9c4-g8tlp\" (UID: \"e153e30f-4f5e-4a38-8efb-6452096e25d7\") " pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.197460 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e153e30f-4f5e-4a38-8efb-6452096e25d7-client-ca-bundle\") pod \"metrics-server-65846dd9c4-g8tlp\" (UID: \"e153e30f-4f5e-4a38-8efb-6452096e25d7\") " pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.198591 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/e153e30f-4f5e-4a38-8efb-6452096e25d7-secret-metrics-server-tls\") pod \"metrics-server-65846dd9c4-g8tlp\" (UID: \"e153e30f-4f5e-4a38-8efb-6452096e25d7\") " pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.204244 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/e153e30f-4f5e-4a38-8efb-6452096e25d7-secret-metrics-client-certs\") pod \"metrics-server-65846dd9c4-g8tlp\" (UID: \"e153e30f-4f5e-4a38-8efb-6452096e25d7\") " pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.214300 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xsqz\" (UniqueName: \"kubernetes.io/projected/e153e30f-4f5e-4a38-8efb-6452096e25d7-kube-api-access-7xsqz\") pod \"metrics-server-65846dd9c4-g8tlp\" (UID: \"e153e30f-4f5e-4a38-8efb-6452096e25d7\") " pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.512978 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.846444 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/monitoring-plugin-6f74c7bb76-pbw94"] Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.848400 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-6f74c7bb76-pbw94" Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.856407 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-monitoring/monitoring-plugin-6c57467cd4-4q6wt"] Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.856694 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-monitoring/monitoring-plugin-6c57467cd4-4q6wt" podUID="0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3" containerName="monitoring-plugin" containerID="cri-o://f1a817deaa3061d3646cc40088ce5badd857d0fbd7404821e4cdff7c61906a1e" gracePeriod=30 Dec 09 09:54:21 crc kubenswrapper[4824]: I1209 09:54:21.867158 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-6f74c7bb76-pbw94"] Dec 09 09:54:22 crc kubenswrapper[4824]: I1209 09:54:22.005601 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/364b1858-6a46-4db4-9d9d-0ec19d54abc9-monitoring-plugin-cert\") pod \"monitoring-plugin-6f74c7bb76-pbw94\" (UID: \"364b1858-6a46-4db4-9d9d-0ec19d54abc9\") " pod="openshift-monitoring/monitoring-plugin-6f74c7bb76-pbw94" Dec 09 09:54:22 crc kubenswrapper[4824]: I1209 09:54:22.107838 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/364b1858-6a46-4db4-9d9d-0ec19d54abc9-monitoring-plugin-cert\") pod \"monitoring-plugin-6f74c7bb76-pbw94\" (UID: \"364b1858-6a46-4db4-9d9d-0ec19d54abc9\") " pod="openshift-monitoring/monitoring-plugin-6f74c7bb76-pbw94" Dec 09 09:54:22 crc kubenswrapper[4824]: I1209 09:54:22.114619 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/364b1858-6a46-4db4-9d9d-0ec19d54abc9-monitoring-plugin-cert\") pod \"monitoring-plugin-6f74c7bb76-pbw94\" (UID: \"364b1858-6a46-4db4-9d9d-0ec19d54abc9\") " pod="openshift-monitoring/monitoring-plugin-6f74c7bb76-pbw94" Dec 09 09:54:22 crc kubenswrapper[4824]: I1209 09:54:22.171382 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-6f74c7bb76-pbw94" Dec 09 09:54:22 crc kubenswrapper[4824]: I1209 09:54:22.302569 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-65846dd9c4-g8tlp"] Dec 09 09:54:22 crc kubenswrapper[4824]: W1209 09:54:22.319115 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode153e30f_4f5e_4a38_8efb_6452096e25d7.slice/crio-42908c6b4fb3bbfa4bb111361e06224fc3c1f821555879885d37d55dbc23327c WatchSource:0}: Error finding container 42908c6b4fb3bbfa4bb111361e06224fc3c1f821555879885d37d55dbc23327c: Status 404 returned error can't find the container with id 42908c6b4fb3bbfa4bb111361e06224fc3c1f821555879885d37d55dbc23327c Dec 09 09:54:22 crc kubenswrapper[4824]: I1209 09:54:22.489946 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_monitoring-plugin-6c57467cd4-4q6wt_0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3/monitoring-plugin/0.log" Dec 09 09:54:22 crc kubenswrapper[4824]: I1209 09:54:22.490052 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-6c57467cd4-4q6wt" Dec 09 09:54:22 crc kubenswrapper[4824]: I1209 09:54:22.603531 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" event={"ID":"e153e30f-4f5e-4a38-8efb-6452096e25d7","Type":"ContainerStarted","Data":"42908c6b4fb3bbfa4bb111361e06224fc3c1f821555879885d37d55dbc23327c"} Dec 09 09:54:22 crc kubenswrapper[4824]: I1209 09:54:22.605622 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_monitoring-plugin-6c57467cd4-4q6wt_0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3/monitoring-plugin/0.log" Dec 09 09:54:22 crc kubenswrapper[4824]: I1209 09:54:22.605671 4824 generic.go:334] "Generic (PLEG): container finished" podID="0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3" containerID="f1a817deaa3061d3646cc40088ce5badd857d0fbd7404821e4cdff7c61906a1e" exitCode=2 Dec 09 09:54:22 crc kubenswrapper[4824]: I1209 09:54:22.605699 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-6c57467cd4-4q6wt" event={"ID":"0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3","Type":"ContainerDied","Data":"f1a817deaa3061d3646cc40088ce5badd857d0fbd7404821e4cdff7c61906a1e"} Dec 09 09:54:22 crc kubenswrapper[4824]: I1209 09:54:22.605726 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-6c57467cd4-4q6wt" event={"ID":"0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3","Type":"ContainerDied","Data":"3a50a62f2dc71b68d4700942b803396c3ba1b8b8963a9e2eb1c34924f3abd31f"} Dec 09 09:54:22 crc kubenswrapper[4824]: I1209 09:54:22.605744 4824 scope.go:117] "RemoveContainer" containerID="f1a817deaa3061d3646cc40088ce5badd857d0fbd7404821e4cdff7c61906a1e" Dec 09 09:54:22 crc kubenswrapper[4824]: I1209 09:54:22.605802 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-6c57467cd4-4q6wt" Dec 09 09:54:22 crc kubenswrapper[4824]: I1209 09:54:22.624186 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3-monitoring-plugin-cert\") pod \"0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3\" (UID: \"0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3\") " Dec 09 09:54:22 crc kubenswrapper[4824]: I1209 09:54:22.627030 4824 scope.go:117] "RemoveContainer" containerID="f1a817deaa3061d3646cc40088ce5badd857d0fbd7404821e4cdff7c61906a1e" Dec 09 09:54:22 crc kubenswrapper[4824]: E1209 09:54:22.627620 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1a817deaa3061d3646cc40088ce5badd857d0fbd7404821e4cdff7c61906a1e\": container with ID starting with f1a817deaa3061d3646cc40088ce5badd857d0fbd7404821e4cdff7c61906a1e not found: ID does not exist" containerID="f1a817deaa3061d3646cc40088ce5badd857d0fbd7404821e4cdff7c61906a1e" Dec 09 09:54:22 crc kubenswrapper[4824]: I1209 09:54:22.627659 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1a817deaa3061d3646cc40088ce5badd857d0fbd7404821e4cdff7c61906a1e"} err="failed to get container status \"f1a817deaa3061d3646cc40088ce5badd857d0fbd7404821e4cdff7c61906a1e\": rpc error: code = NotFound desc = could not find container \"f1a817deaa3061d3646cc40088ce5badd857d0fbd7404821e4cdff7c61906a1e\": container with ID starting with f1a817deaa3061d3646cc40088ce5badd857d0fbd7404821e4cdff7c61906a1e not found: ID does not exist" Dec 09 09:54:22 crc kubenswrapper[4824]: I1209 09:54:22.629907 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3-monitoring-plugin-cert" (OuterVolumeSpecName: "monitoring-plugin-cert") pod "0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3" (UID: "0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3"). InnerVolumeSpecName "monitoring-plugin-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:54:22 crc kubenswrapper[4824]: I1209 09:54:22.688443 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-6f74c7bb76-pbw94"] Dec 09 09:54:22 crc kubenswrapper[4824]: W1209 09:54:22.693966 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod364b1858_6a46_4db4_9d9d_0ec19d54abc9.slice/crio-036c7f01a860312a93531c3d0b46054935ec8997f7c7b52db379faa0ff7bb7e8 WatchSource:0}: Error finding container 036c7f01a860312a93531c3d0b46054935ec8997f7c7b52db379faa0ff7bb7e8: Status 404 returned error can't find the container with id 036c7f01a860312a93531c3d0b46054935ec8997f7c7b52db379faa0ff7bb7e8 Dec 09 09:54:22 crc kubenswrapper[4824]: I1209 09:54:22.745740 4824 reconciler_common.go:293] "Volume detached for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3-monitoring-plugin-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:54:22 crc kubenswrapper[4824]: I1209 09:54:22.938549 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-monitoring/monitoring-plugin-6c57467cd4-4q6wt"] Dec 09 09:54:22 crc kubenswrapper[4824]: I1209 09:54:22.943790 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-monitoring/monitoring-plugin-6c57467cd4-4q6wt"] Dec 09 09:54:23 crc kubenswrapper[4824]: I1209 09:54:23.617303 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-6f74c7bb76-pbw94" event={"ID":"364b1858-6a46-4db4-9d9d-0ec19d54abc9","Type":"ContainerStarted","Data":"cb48d4518a8e41132a38871227763b3ea9c87948149ba8af0b0b9a01ec648610"} Dec 09 09:54:23 crc kubenswrapper[4824]: I1209 09:54:23.617901 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/monitoring-plugin-6f74c7bb76-pbw94" Dec 09 09:54:23 crc kubenswrapper[4824]: I1209 09:54:23.617921 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-6f74c7bb76-pbw94" event={"ID":"364b1858-6a46-4db4-9d9d-0ec19d54abc9","Type":"ContainerStarted","Data":"036c7f01a860312a93531c3d0b46054935ec8997f7c7b52db379faa0ff7bb7e8"} Dec 09 09:54:23 crc kubenswrapper[4824]: I1209 09:54:23.620900 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" event={"ID":"e153e30f-4f5e-4a38-8efb-6452096e25d7","Type":"ContainerStarted","Data":"7da26121a429a48f86bdeaadb323c671e73ba9993fa5bd40f3fcca78a5bf3769"} Dec 09 09:54:23 crc kubenswrapper[4824]: I1209 09:54:23.627756 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/monitoring-plugin-6f74c7bb76-pbw94" Dec 09 09:54:23 crc kubenswrapper[4824]: I1209 09:54:23.637097 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/monitoring-plugin-6f74c7bb76-pbw94" podStartSLOduration=2.637071869 podStartE2EDuration="2.637071869s" podCreationTimestamp="2025-12-09 09:54:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:54:23.633943711 +0000 UTC m=+419.968448388" watchObservedRunningTime="2025-12-09 09:54:23.637071869 +0000 UTC m=+419.971576536" Dec 09 09:54:23 crc kubenswrapper[4824]: I1209 09:54:23.685762 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" podStartSLOduration=3.685739212 podStartE2EDuration="3.685739212s" podCreationTimestamp="2025-12-09 09:54:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:54:23.683594224 +0000 UTC m=+420.018098891" watchObservedRunningTime="2025-12-09 09:54:23.685739212 +0000 UTC m=+420.020243879" Dec 09 09:54:23 crc kubenswrapper[4824]: I1209 09:54:23.919947 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3" path="/var/lib/kubelet/pods/0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3/volumes" Dec 09 09:54:25 crc kubenswrapper[4824]: I1209 09:54:25.705599 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-79d58c8b9c-dbzmt"] Dec 09 09:54:25 crc kubenswrapper[4824]: E1209 09:54:25.706394 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3" containerName="monitoring-plugin" Dec 09 09:54:25 crc kubenswrapper[4824]: I1209 09:54:25.706412 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3" containerName="monitoring-plugin" Dec 09 09:54:25 crc kubenswrapper[4824]: I1209 09:54:25.706556 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c4a05b0-24d6-4ffd-a2d1-8e0f9af049f3" containerName="monitoring-plugin" Dec 09 09:54:25 crc kubenswrapper[4824]: I1209 09:54:25.707219 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:25 crc kubenswrapper[4824]: I1209 09:54:25.736414 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-79d58c8b9c-dbzmt"] Dec 09 09:54:25 crc kubenswrapper[4824]: I1209 09:54:25.906624 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/065d385c-89bf-409f-822e-58652c09a9a3-console-oauth-config\") pod \"console-79d58c8b9c-dbzmt\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:25 crc kubenswrapper[4824]: I1209 09:54:25.907074 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/065d385c-89bf-409f-822e-58652c09a9a3-trusted-ca-bundle\") pod \"console-79d58c8b9c-dbzmt\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:25 crc kubenswrapper[4824]: I1209 09:54:25.907121 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/065d385c-89bf-409f-822e-58652c09a9a3-service-ca\") pod \"console-79d58c8b9c-dbzmt\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:25 crc kubenswrapper[4824]: I1209 09:54:25.907165 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/065d385c-89bf-409f-822e-58652c09a9a3-console-serving-cert\") pod \"console-79d58c8b9c-dbzmt\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:25 crc kubenswrapper[4824]: I1209 09:54:25.907187 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/065d385c-89bf-409f-822e-58652c09a9a3-console-config\") pod \"console-79d58c8b9c-dbzmt\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:25 crc kubenswrapper[4824]: I1209 09:54:25.907212 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl5v8\" (UniqueName: \"kubernetes.io/projected/065d385c-89bf-409f-822e-58652c09a9a3-kube-api-access-fl5v8\") pod \"console-79d58c8b9c-dbzmt\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:25 crc kubenswrapper[4824]: I1209 09:54:25.907247 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/065d385c-89bf-409f-822e-58652c09a9a3-oauth-serving-cert\") pod \"console-79d58c8b9c-dbzmt\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:26 crc kubenswrapper[4824]: I1209 09:54:26.008548 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/065d385c-89bf-409f-822e-58652c09a9a3-console-oauth-config\") pod \"console-79d58c8b9c-dbzmt\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:26 crc kubenswrapper[4824]: I1209 09:54:26.008626 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/065d385c-89bf-409f-822e-58652c09a9a3-trusted-ca-bundle\") pod \"console-79d58c8b9c-dbzmt\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:26 crc kubenswrapper[4824]: I1209 09:54:26.008667 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/065d385c-89bf-409f-822e-58652c09a9a3-service-ca\") pod \"console-79d58c8b9c-dbzmt\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:26 crc kubenswrapper[4824]: I1209 09:54:26.008705 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/065d385c-89bf-409f-822e-58652c09a9a3-console-serving-cert\") pod \"console-79d58c8b9c-dbzmt\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:26 crc kubenswrapper[4824]: I1209 09:54:26.008733 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/065d385c-89bf-409f-822e-58652c09a9a3-console-config\") pod \"console-79d58c8b9c-dbzmt\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:26 crc kubenswrapper[4824]: I1209 09:54:26.008776 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl5v8\" (UniqueName: \"kubernetes.io/projected/065d385c-89bf-409f-822e-58652c09a9a3-kube-api-access-fl5v8\") pod \"console-79d58c8b9c-dbzmt\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:26 crc kubenswrapper[4824]: I1209 09:54:26.008828 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/065d385c-89bf-409f-822e-58652c09a9a3-oauth-serving-cert\") pod \"console-79d58c8b9c-dbzmt\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:26 crc kubenswrapper[4824]: I1209 09:54:26.010182 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/065d385c-89bf-409f-822e-58652c09a9a3-oauth-serving-cert\") pod \"console-79d58c8b9c-dbzmt\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:26 crc kubenswrapper[4824]: I1209 09:54:26.010580 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/065d385c-89bf-409f-822e-58652c09a9a3-service-ca\") pod \"console-79d58c8b9c-dbzmt\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:26 crc kubenswrapper[4824]: I1209 09:54:26.010614 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/065d385c-89bf-409f-822e-58652c09a9a3-trusted-ca-bundle\") pod \"console-79d58c8b9c-dbzmt\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:26 crc kubenswrapper[4824]: I1209 09:54:26.010599 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/065d385c-89bf-409f-822e-58652c09a9a3-console-config\") pod \"console-79d58c8b9c-dbzmt\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:26 crc kubenswrapper[4824]: I1209 09:54:26.017508 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/065d385c-89bf-409f-822e-58652c09a9a3-console-oauth-config\") pod \"console-79d58c8b9c-dbzmt\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:26 crc kubenswrapper[4824]: I1209 09:54:26.017640 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/065d385c-89bf-409f-822e-58652c09a9a3-console-serving-cert\") pod \"console-79d58c8b9c-dbzmt\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:26 crc kubenswrapper[4824]: I1209 09:54:26.031245 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl5v8\" (UniqueName: \"kubernetes.io/projected/065d385c-89bf-409f-822e-58652c09a9a3-kube-api-access-fl5v8\") pod \"console-79d58c8b9c-dbzmt\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:26 crc kubenswrapper[4824]: I1209 09:54:26.325662 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:26 crc kubenswrapper[4824]: I1209 09:54:26.798735 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-79d58c8b9c-dbzmt"] Dec 09 09:54:26 crc kubenswrapper[4824]: W1209 09:54:26.805192 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod065d385c_89bf_409f_822e_58652c09a9a3.slice/crio-7e0db67401027fc28ff91cf556926cd41abb94b9bff90553a0f5f023c6631532 WatchSource:0}: Error finding container 7e0db67401027fc28ff91cf556926cd41abb94b9bff90553a0f5f023c6631532: Status 404 returned error can't find the container with id 7e0db67401027fc28ff91cf556926cd41abb94b9bff90553a0f5f023c6631532 Dec 09 09:54:27 crc kubenswrapper[4824]: I1209 09:54:27.720223 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-79d58c8b9c-dbzmt" event={"ID":"065d385c-89bf-409f-822e-58652c09a9a3","Type":"ContainerStarted","Data":"d675c1d6889d781b51f200d99f17df80a14f3b02b3443ab8758e818dd96307f6"} Dec 09 09:54:27 crc kubenswrapper[4824]: I1209 09:54:27.720744 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-79d58c8b9c-dbzmt" event={"ID":"065d385c-89bf-409f-822e-58652c09a9a3","Type":"ContainerStarted","Data":"7e0db67401027fc28ff91cf556926cd41abb94b9bff90553a0f5f023c6631532"} Dec 09 09:54:27 crc kubenswrapper[4824]: I1209 09:54:27.769541 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-79d58c8b9c-dbzmt" podStartSLOduration=2.769516721 podStartE2EDuration="2.769516721s" podCreationTimestamp="2025-12-09 09:54:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:54:27.764050219 +0000 UTC m=+424.098554886" watchObservedRunningTime="2025-12-09 09:54:27.769516721 +0000 UTC m=+424.104021388" Dec 09 09:54:32 crc kubenswrapper[4824]: I1209 09:54:32.861929 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 09:54:32 crc kubenswrapper[4824]: I1209 09:54:32.862734 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 09:54:32 crc kubenswrapper[4824]: I1209 09:54:32.862860 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 09:54:32 crc kubenswrapper[4824]: I1209 09:54:32.863921 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ea7720fbe70782cdf3ecb70ae54c936954803442a5c7d2c6bc34bd0e635bfbfe"} pod="openshift-machine-config-operator/machine-config-daemon-dth8x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 09:54:32 crc kubenswrapper[4824]: I1209 09:54:32.863991 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" containerID="cri-o://ea7720fbe70782cdf3ecb70ae54c936954803442a5c7d2c6bc34bd0e635bfbfe" gracePeriod=600 Dec 09 09:54:33 crc kubenswrapper[4824]: I1209 09:54:33.772234 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerID="ea7720fbe70782cdf3ecb70ae54c936954803442a5c7d2c6bc34bd0e635bfbfe" exitCode=0 Dec 09 09:54:33 crc kubenswrapper[4824]: I1209 09:54:33.772307 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerDied","Data":"ea7720fbe70782cdf3ecb70ae54c936954803442a5c7d2c6bc34bd0e635bfbfe"} Dec 09 09:54:33 crc kubenswrapper[4824]: I1209 09:54:33.772816 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerStarted","Data":"b9242f271427aa606711a2a58ca0863fdbf70e4a738de9f900b42d848af928bc"} Dec 09 09:54:33 crc kubenswrapper[4824]: I1209 09:54:33.772868 4824 scope.go:117] "RemoveContainer" containerID="314663cd69c33a83e02c227178f74fd4f1dcedd4e4287889e9a7b644ab345671" Dec 09 09:54:36 crc kubenswrapper[4824]: I1209 09:54:36.326414 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:36 crc kubenswrapper[4824]: I1209 09:54:36.327225 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:36 crc kubenswrapper[4824]: I1209 09:54:36.332537 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:36 crc kubenswrapper[4824]: I1209 09:54:36.797198 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 09:54:36 crc kubenswrapper[4824]: I1209 09:54:36.875007 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-684fc64449-59s5g"] Dec 09 09:54:41 crc kubenswrapper[4824]: I1209 09:54:41.513354 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:54:41 crc kubenswrapper[4824]: I1209 09:54:41.514230 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:55:01 crc kubenswrapper[4824]: I1209 09:55:01.519942 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:55:01 crc kubenswrapper[4824]: I1209 09:55:01.526925 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 09:55:01 crc kubenswrapper[4824]: I1209 09:55:01.927323 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-684fc64449-59s5g" podUID="536256e5-73d0-4de7-8f6d-4e5d449b84ef" containerName="console" containerID="cri-o://2b8b31e3c2c987595769cb00f48cfbd5cf000e9273ee2990febb6749308b007b" gracePeriod=15 Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.493664 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-684fc64449-59s5g_536256e5-73d0-4de7-8f6d-4e5d449b84ef/console/0.log" Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.494142 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.539029 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/536256e5-73d0-4de7-8f6d-4e5d449b84ef-console-oauth-config\") pod \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.539142 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/536256e5-73d0-4de7-8f6d-4e5d449b84ef-console-serving-cert\") pod \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.539225 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9c22f\" (UniqueName: \"kubernetes.io/projected/536256e5-73d0-4de7-8f6d-4e5d449b84ef-kube-api-access-9c22f\") pod \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.539291 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/536256e5-73d0-4de7-8f6d-4e5d449b84ef-trusted-ca-bundle\") pod \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.539322 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/536256e5-73d0-4de7-8f6d-4e5d449b84ef-oauth-serving-cert\") pod \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.539375 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/536256e5-73d0-4de7-8f6d-4e5d449b84ef-console-config\") pod \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.539425 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/536256e5-73d0-4de7-8f6d-4e5d449b84ef-service-ca\") pod \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\" (UID: \"536256e5-73d0-4de7-8f6d-4e5d449b84ef\") " Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.541314 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/536256e5-73d0-4de7-8f6d-4e5d449b84ef-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "536256e5-73d0-4de7-8f6d-4e5d449b84ef" (UID: "536256e5-73d0-4de7-8f6d-4e5d449b84ef"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.541374 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/536256e5-73d0-4de7-8f6d-4e5d449b84ef-console-config" (OuterVolumeSpecName: "console-config") pod "536256e5-73d0-4de7-8f6d-4e5d449b84ef" (UID: "536256e5-73d0-4de7-8f6d-4e5d449b84ef"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.541406 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/536256e5-73d0-4de7-8f6d-4e5d449b84ef-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "536256e5-73d0-4de7-8f6d-4e5d449b84ef" (UID: "536256e5-73d0-4de7-8f6d-4e5d449b84ef"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.542746 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/536256e5-73d0-4de7-8f6d-4e5d449b84ef-service-ca" (OuterVolumeSpecName: "service-ca") pod "536256e5-73d0-4de7-8f6d-4e5d449b84ef" (UID: "536256e5-73d0-4de7-8f6d-4e5d449b84ef"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.547590 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/536256e5-73d0-4de7-8f6d-4e5d449b84ef-kube-api-access-9c22f" (OuterVolumeSpecName: "kube-api-access-9c22f") pod "536256e5-73d0-4de7-8f6d-4e5d449b84ef" (UID: "536256e5-73d0-4de7-8f6d-4e5d449b84ef"). InnerVolumeSpecName "kube-api-access-9c22f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.547613 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/536256e5-73d0-4de7-8f6d-4e5d449b84ef-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "536256e5-73d0-4de7-8f6d-4e5d449b84ef" (UID: "536256e5-73d0-4de7-8f6d-4e5d449b84ef"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.549825 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/536256e5-73d0-4de7-8f6d-4e5d449b84ef-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "536256e5-73d0-4de7-8f6d-4e5d449b84ef" (UID: "536256e5-73d0-4de7-8f6d-4e5d449b84ef"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.640925 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/536256e5-73d0-4de7-8f6d-4e5d449b84ef-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.640987 4824 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/536256e5-73d0-4de7-8f6d-4e5d449b84ef-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.641004 4824 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/536256e5-73d0-4de7-8f6d-4e5d449b84ef-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.641049 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9c22f\" (UniqueName: \"kubernetes.io/projected/536256e5-73d0-4de7-8f6d-4e5d449b84ef-kube-api-access-9c22f\") on node \"crc\" DevicePath \"\"" Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.641060 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/536256e5-73d0-4de7-8f6d-4e5d449b84ef-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.641073 4824 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/536256e5-73d0-4de7-8f6d-4e5d449b84ef-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.641084 4824 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/536256e5-73d0-4de7-8f6d-4e5d449b84ef-console-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.976659 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-684fc64449-59s5g_536256e5-73d0-4de7-8f6d-4e5d449b84ef/console/0.log" Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.977093 4824 generic.go:334] "Generic (PLEG): container finished" podID="536256e5-73d0-4de7-8f6d-4e5d449b84ef" containerID="2b8b31e3c2c987595769cb00f48cfbd5cf000e9273ee2990febb6749308b007b" exitCode=2 Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.977203 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-684fc64449-59s5g" event={"ID":"536256e5-73d0-4de7-8f6d-4e5d449b84ef","Type":"ContainerDied","Data":"2b8b31e3c2c987595769cb00f48cfbd5cf000e9273ee2990febb6749308b007b"} Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.977295 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-684fc64449-59s5g" event={"ID":"536256e5-73d0-4de7-8f6d-4e5d449b84ef","Type":"ContainerDied","Data":"1bba57fbb6e3ab84e8fb5d4c7d6e01d3675150d3a55100691f2b7e50f6012582"} Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.977324 4824 scope.go:117] "RemoveContainer" containerID="2b8b31e3c2c987595769cb00f48cfbd5cf000e9273ee2990febb6749308b007b" Dec 09 09:55:02 crc kubenswrapper[4824]: I1209 09:55:02.979341 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-684fc64449-59s5g" Dec 09 09:55:03 crc kubenswrapper[4824]: I1209 09:55:03.001967 4824 scope.go:117] "RemoveContainer" containerID="2b8b31e3c2c987595769cb00f48cfbd5cf000e9273ee2990febb6749308b007b" Dec 09 09:55:03 crc kubenswrapper[4824]: E1209 09:55:03.005060 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b8b31e3c2c987595769cb00f48cfbd5cf000e9273ee2990febb6749308b007b\": container with ID starting with 2b8b31e3c2c987595769cb00f48cfbd5cf000e9273ee2990febb6749308b007b not found: ID does not exist" containerID="2b8b31e3c2c987595769cb00f48cfbd5cf000e9273ee2990febb6749308b007b" Dec 09 09:55:03 crc kubenswrapper[4824]: I1209 09:55:03.005133 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b8b31e3c2c987595769cb00f48cfbd5cf000e9273ee2990febb6749308b007b"} err="failed to get container status \"2b8b31e3c2c987595769cb00f48cfbd5cf000e9273ee2990febb6749308b007b\": rpc error: code = NotFound desc = could not find container \"2b8b31e3c2c987595769cb00f48cfbd5cf000e9273ee2990febb6749308b007b\": container with ID starting with 2b8b31e3c2c987595769cb00f48cfbd5cf000e9273ee2990febb6749308b007b not found: ID does not exist" Dec 09 09:55:03 crc kubenswrapper[4824]: I1209 09:55:03.023625 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-684fc64449-59s5g"] Dec 09 09:55:03 crc kubenswrapper[4824]: I1209 09:55:03.030125 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-684fc64449-59s5g"] Dec 09 09:55:03 crc kubenswrapper[4824]: I1209 09:55:03.919940 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="536256e5-73d0-4de7-8f6d-4e5d449b84ef" path="/var/lib/kubelet/pods/536256e5-73d0-4de7-8f6d-4e5d449b84ef/volumes" Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.279357 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.368535 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa0314b-a177-4589-bc32-f9caf6f8f724-client-ca-bundle\") pod \"9aa0314b-a177-4589-bc32-f9caf6f8f724\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.368641 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/9aa0314b-a177-4589-bc32-f9caf6f8f724-metrics-server-audit-profiles\") pod \"9aa0314b-a177-4589-bc32-f9caf6f8f724\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.368683 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9aa0314b-a177-4589-bc32-f9caf6f8f724-configmap-kubelet-serving-ca-bundle\") pod \"9aa0314b-a177-4589-bc32-f9caf6f8f724\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.368722 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/9aa0314b-a177-4589-bc32-f9caf6f8f724-secret-metrics-server-tls\") pod \"9aa0314b-a177-4589-bc32-f9caf6f8f724\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.368769 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/9aa0314b-a177-4589-bc32-f9caf6f8f724-audit-log\") pod \"9aa0314b-a177-4589-bc32-f9caf6f8f724\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.368878 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/9aa0314b-a177-4589-bc32-f9caf6f8f724-secret-metrics-client-certs\") pod \"9aa0314b-a177-4589-bc32-f9caf6f8f724\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.368955 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfprf\" (UniqueName: \"kubernetes.io/projected/9aa0314b-a177-4589-bc32-f9caf6f8f724-kube-api-access-gfprf\") pod \"9aa0314b-a177-4589-bc32-f9caf6f8f724\" (UID: \"9aa0314b-a177-4589-bc32-f9caf6f8f724\") " Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.369665 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aa0314b-a177-4589-bc32-f9caf6f8f724-configmap-kubelet-serving-ca-bundle" (OuterVolumeSpecName: "configmap-kubelet-serving-ca-bundle") pod "9aa0314b-a177-4589-bc32-f9caf6f8f724" (UID: "9aa0314b-a177-4589-bc32-f9caf6f8f724"). InnerVolumeSpecName "configmap-kubelet-serving-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.369676 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aa0314b-a177-4589-bc32-f9caf6f8f724-metrics-server-audit-profiles" (OuterVolumeSpecName: "metrics-server-audit-profiles") pod "9aa0314b-a177-4589-bc32-f9caf6f8f724" (UID: "9aa0314b-a177-4589-bc32-f9caf6f8f724"). InnerVolumeSpecName "metrics-server-audit-profiles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.370250 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9aa0314b-a177-4589-bc32-f9caf6f8f724-audit-log" (OuterVolumeSpecName: "audit-log") pod "9aa0314b-a177-4589-bc32-f9caf6f8f724" (UID: "9aa0314b-a177-4589-bc32-f9caf6f8f724"). InnerVolumeSpecName "audit-log". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.374942 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aa0314b-a177-4589-bc32-f9caf6f8f724-secret-metrics-server-tls" (OuterVolumeSpecName: "secret-metrics-server-tls") pod "9aa0314b-a177-4589-bc32-f9caf6f8f724" (UID: "9aa0314b-a177-4589-bc32-f9caf6f8f724"). InnerVolumeSpecName "secret-metrics-server-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.375062 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aa0314b-a177-4589-bc32-f9caf6f8f724-client-ca-bundle" (OuterVolumeSpecName: "client-ca-bundle") pod "9aa0314b-a177-4589-bc32-f9caf6f8f724" (UID: "9aa0314b-a177-4589-bc32-f9caf6f8f724"). InnerVolumeSpecName "client-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.375088 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aa0314b-a177-4589-bc32-f9caf6f8f724-kube-api-access-gfprf" (OuterVolumeSpecName: "kube-api-access-gfprf") pod "9aa0314b-a177-4589-bc32-f9caf6f8f724" (UID: "9aa0314b-a177-4589-bc32-f9caf6f8f724"). InnerVolumeSpecName "kube-api-access-gfprf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.375420 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aa0314b-a177-4589-bc32-f9caf6f8f724-secret-metrics-client-certs" (OuterVolumeSpecName: "secret-metrics-client-certs") pod "9aa0314b-a177-4589-bc32-f9caf6f8f724" (UID: "9aa0314b-a177-4589-bc32-f9caf6f8f724"). InnerVolumeSpecName "secret-metrics-client-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.470859 4824 reconciler_common.go:293] "Volume detached for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9aa0314b-a177-4589-bc32-f9caf6f8f724-configmap-kubelet-serving-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.470939 4824 reconciler_common.go:293] "Volume detached for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/9aa0314b-a177-4589-bc32-f9caf6f8f724-secret-metrics-server-tls\") on node \"crc\" DevicePath \"\"" Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.470955 4824 reconciler_common.go:293] "Volume detached for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/9aa0314b-a177-4589-bc32-f9caf6f8f724-audit-log\") on node \"crc\" DevicePath \"\"" Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.470969 4824 reconciler_common.go:293] "Volume detached for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/9aa0314b-a177-4589-bc32-f9caf6f8f724-secret-metrics-client-certs\") on node \"crc\" DevicePath \"\"" Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.470989 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfprf\" (UniqueName: \"kubernetes.io/projected/9aa0314b-a177-4589-bc32-f9caf6f8f724-kube-api-access-gfprf\") on node \"crc\" DevicePath \"\"" Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.471001 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa0314b-a177-4589-bc32-f9caf6f8f724-client-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.471015 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/9aa0314b-a177-4589-bc32-f9caf6f8f724-metrics-server-audit-profiles\") on node \"crc\" DevicePath \"\"" Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.753869 4824 generic.go:334] "Generic (PLEG): container finished" podID="9aa0314b-a177-4589-bc32-f9caf6f8f724" containerID="c3e8d0870737ba4e5bbc0d028449d281500635ecaf1a9525b3d108cd03a45b6c" exitCode=0 Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.754499 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" event={"ID":"9aa0314b-a177-4589-bc32-f9caf6f8f724","Type":"ContainerDied","Data":"c3e8d0870737ba4e5bbc0d028449d281500635ecaf1a9525b3d108cd03a45b6c"} Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.754603 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" event={"ID":"9aa0314b-a177-4589-bc32-f9caf6f8f724","Type":"ContainerDied","Data":"509989dd226b9819afb3d9917ad61603277049db1578fcba6fefba79ce205ab4"} Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.754728 4824 scope.go:117] "RemoveContainer" containerID="c3e8d0870737ba4e5bbc0d028449d281500635ecaf1a9525b3d108cd03a45b6c" Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.754980 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-6448d488d4-xnf9s" Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.799377 4824 scope.go:117] "RemoveContainer" containerID="c3e8d0870737ba4e5bbc0d028449d281500635ecaf1a9525b3d108cd03a45b6c" Dec 09 09:56:51 crc kubenswrapper[4824]: E1209 09:56:51.812990 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3e8d0870737ba4e5bbc0d028449d281500635ecaf1a9525b3d108cd03a45b6c\": container with ID starting with c3e8d0870737ba4e5bbc0d028449d281500635ecaf1a9525b3d108cd03a45b6c not found: ID does not exist" containerID="c3e8d0870737ba4e5bbc0d028449d281500635ecaf1a9525b3d108cd03a45b6c" Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.813067 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3e8d0870737ba4e5bbc0d028449d281500635ecaf1a9525b3d108cd03a45b6c"} err="failed to get container status \"c3e8d0870737ba4e5bbc0d028449d281500635ecaf1a9525b3d108cd03a45b6c\": rpc error: code = NotFound desc = could not find container \"c3e8d0870737ba4e5bbc0d028449d281500635ecaf1a9525b3d108cd03a45b6c\": container with ID starting with c3e8d0870737ba4e5bbc0d028449d281500635ecaf1a9525b3d108cd03a45b6c not found: ID does not exist" Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.821835 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-monitoring/metrics-server-6448d488d4-xnf9s"] Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.832850 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-monitoring/metrics-server-6448d488d4-xnf9s"] Dec 09 09:56:51 crc kubenswrapper[4824]: I1209 09:56:51.920154 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9aa0314b-a177-4589-bc32-f9caf6f8f724" path="/var/lib/kubelet/pods/9aa0314b-a177-4589-bc32-f9caf6f8f724/volumes" Dec 09 09:56:58 crc kubenswrapper[4824]: I1209 09:56:58.303006 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8"] Dec 09 09:56:58 crc kubenswrapper[4824]: E1209 09:56:58.303939 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="536256e5-73d0-4de7-8f6d-4e5d449b84ef" containerName="console" Dec 09 09:56:58 crc kubenswrapper[4824]: I1209 09:56:58.303954 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="536256e5-73d0-4de7-8f6d-4e5d449b84ef" containerName="console" Dec 09 09:56:58 crc kubenswrapper[4824]: E1209 09:56:58.303968 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aa0314b-a177-4589-bc32-f9caf6f8f724" containerName="metrics-server" Dec 09 09:56:58 crc kubenswrapper[4824]: I1209 09:56:58.303975 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aa0314b-a177-4589-bc32-f9caf6f8f724" containerName="metrics-server" Dec 09 09:56:58 crc kubenswrapper[4824]: I1209 09:56:58.304094 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aa0314b-a177-4589-bc32-f9caf6f8f724" containerName="metrics-server" Dec 09 09:56:58 crc kubenswrapper[4824]: I1209 09:56:58.304104 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="536256e5-73d0-4de7-8f6d-4e5d449b84ef" containerName="console" Dec 09 09:56:58 crc kubenswrapper[4824]: I1209 09:56:58.304994 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8" Dec 09 09:56:58 crc kubenswrapper[4824]: I1209 09:56:58.312365 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 09 09:56:58 crc kubenswrapper[4824]: I1209 09:56:58.317699 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8"] Dec 09 09:56:58 crc kubenswrapper[4824]: I1209 09:56:58.368387 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tptbv\" (UniqueName: \"kubernetes.io/projected/d28511a4-2458-434b-85ad-98d2253d4d7b-kube-api-access-tptbv\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8\" (UID: \"d28511a4-2458-434b-85ad-98d2253d4d7b\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8" Dec 09 09:56:58 crc kubenswrapper[4824]: I1209 09:56:58.368475 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d28511a4-2458-434b-85ad-98d2253d4d7b-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8\" (UID: \"d28511a4-2458-434b-85ad-98d2253d4d7b\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8" Dec 09 09:56:58 crc kubenswrapper[4824]: I1209 09:56:58.368725 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d28511a4-2458-434b-85ad-98d2253d4d7b-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8\" (UID: \"d28511a4-2458-434b-85ad-98d2253d4d7b\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8" Dec 09 09:56:58 crc kubenswrapper[4824]: I1209 09:56:58.470932 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d28511a4-2458-434b-85ad-98d2253d4d7b-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8\" (UID: \"d28511a4-2458-434b-85ad-98d2253d4d7b\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8" Dec 09 09:56:58 crc kubenswrapper[4824]: I1209 09:56:58.471097 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tptbv\" (UniqueName: \"kubernetes.io/projected/d28511a4-2458-434b-85ad-98d2253d4d7b-kube-api-access-tptbv\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8\" (UID: \"d28511a4-2458-434b-85ad-98d2253d4d7b\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8" Dec 09 09:56:58 crc kubenswrapper[4824]: I1209 09:56:58.471186 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d28511a4-2458-434b-85ad-98d2253d4d7b-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8\" (UID: \"d28511a4-2458-434b-85ad-98d2253d4d7b\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8" Dec 09 09:56:58 crc kubenswrapper[4824]: I1209 09:56:58.473345 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d28511a4-2458-434b-85ad-98d2253d4d7b-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8\" (UID: \"d28511a4-2458-434b-85ad-98d2253d4d7b\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8" Dec 09 09:56:58 crc kubenswrapper[4824]: I1209 09:56:58.473633 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d28511a4-2458-434b-85ad-98d2253d4d7b-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8\" (UID: \"d28511a4-2458-434b-85ad-98d2253d4d7b\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8" Dec 09 09:56:58 crc kubenswrapper[4824]: I1209 09:56:58.497758 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tptbv\" (UniqueName: \"kubernetes.io/projected/d28511a4-2458-434b-85ad-98d2253d4d7b-kube-api-access-tptbv\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8\" (UID: \"d28511a4-2458-434b-85ad-98d2253d4d7b\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8" Dec 09 09:56:58 crc kubenswrapper[4824]: I1209 09:56:58.625512 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8" Dec 09 09:56:58 crc kubenswrapper[4824]: I1209 09:56:58.911728 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8"] Dec 09 09:56:59 crc kubenswrapper[4824]: I1209 09:56:59.859375 4824 generic.go:334] "Generic (PLEG): container finished" podID="d28511a4-2458-434b-85ad-98d2253d4d7b" containerID="fb09ce50334f8e23e1f04572f4d4c9debe78a6a267dc6211992bf78044b8f701" exitCode=0 Dec 09 09:56:59 crc kubenswrapper[4824]: I1209 09:56:59.859454 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8" event={"ID":"d28511a4-2458-434b-85ad-98d2253d4d7b","Type":"ContainerDied","Data":"fb09ce50334f8e23e1f04572f4d4c9debe78a6a267dc6211992bf78044b8f701"} Dec 09 09:56:59 crc kubenswrapper[4824]: I1209 09:56:59.859969 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8" event={"ID":"d28511a4-2458-434b-85ad-98d2253d4d7b","Type":"ContainerStarted","Data":"21065b9b332c0d268044dace87b98390c310d6ba7ca63f0aa6cac94bdba87b00"} Dec 09 09:56:59 crc kubenswrapper[4824]: I1209 09:56:59.862573 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 09:57:01 crc kubenswrapper[4824]: I1209 09:57:01.883775 4824 generic.go:334] "Generic (PLEG): container finished" podID="d28511a4-2458-434b-85ad-98d2253d4d7b" containerID="30540428f1bdec89e31d7f06a583080f9c220884d1560e96e6170bb6cb27506d" exitCode=0 Dec 09 09:57:01 crc kubenswrapper[4824]: I1209 09:57:01.883829 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8" event={"ID":"d28511a4-2458-434b-85ad-98d2253d4d7b","Type":"ContainerDied","Data":"30540428f1bdec89e31d7f06a583080f9c220884d1560e96e6170bb6cb27506d"} Dec 09 09:57:02 crc kubenswrapper[4824]: I1209 09:57:02.861118 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 09:57:02 crc kubenswrapper[4824]: I1209 09:57:02.861631 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 09:57:02 crc kubenswrapper[4824]: I1209 09:57:02.899545 4824 generic.go:334] "Generic (PLEG): container finished" podID="d28511a4-2458-434b-85ad-98d2253d4d7b" containerID="9982cd4d97a6568a302648a3d2be18e9eff1620b1542e86f05dcb5bcb9d44d1c" exitCode=0 Dec 09 09:57:02 crc kubenswrapper[4824]: I1209 09:57:02.899628 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8" event={"ID":"d28511a4-2458-434b-85ad-98d2253d4d7b","Type":"ContainerDied","Data":"9982cd4d97a6568a302648a3d2be18e9eff1620b1542e86f05dcb5bcb9d44d1c"} Dec 09 09:57:04 crc kubenswrapper[4824]: I1209 09:57:04.182681 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8" Dec 09 09:57:04 crc kubenswrapper[4824]: I1209 09:57:04.274900 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tptbv\" (UniqueName: \"kubernetes.io/projected/d28511a4-2458-434b-85ad-98d2253d4d7b-kube-api-access-tptbv\") pod \"d28511a4-2458-434b-85ad-98d2253d4d7b\" (UID: \"d28511a4-2458-434b-85ad-98d2253d4d7b\") " Dec 09 09:57:04 crc kubenswrapper[4824]: I1209 09:57:04.275007 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d28511a4-2458-434b-85ad-98d2253d4d7b-util\") pod \"d28511a4-2458-434b-85ad-98d2253d4d7b\" (UID: \"d28511a4-2458-434b-85ad-98d2253d4d7b\") " Dec 09 09:57:04 crc kubenswrapper[4824]: I1209 09:57:04.275087 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d28511a4-2458-434b-85ad-98d2253d4d7b-bundle\") pod \"d28511a4-2458-434b-85ad-98d2253d4d7b\" (UID: \"d28511a4-2458-434b-85ad-98d2253d4d7b\") " Dec 09 09:57:04 crc kubenswrapper[4824]: I1209 09:57:04.279297 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d28511a4-2458-434b-85ad-98d2253d4d7b-bundle" (OuterVolumeSpecName: "bundle") pod "d28511a4-2458-434b-85ad-98d2253d4d7b" (UID: "d28511a4-2458-434b-85ad-98d2253d4d7b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:57:04 crc kubenswrapper[4824]: I1209 09:57:04.283831 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d28511a4-2458-434b-85ad-98d2253d4d7b-kube-api-access-tptbv" (OuterVolumeSpecName: "kube-api-access-tptbv") pod "d28511a4-2458-434b-85ad-98d2253d4d7b" (UID: "d28511a4-2458-434b-85ad-98d2253d4d7b"). InnerVolumeSpecName "kube-api-access-tptbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:57:04 crc kubenswrapper[4824]: I1209 09:57:04.291968 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d28511a4-2458-434b-85ad-98d2253d4d7b-util" (OuterVolumeSpecName: "util") pod "d28511a4-2458-434b-85ad-98d2253d4d7b" (UID: "d28511a4-2458-434b-85ad-98d2253d4d7b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:57:04 crc kubenswrapper[4824]: I1209 09:57:04.376796 4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d28511a4-2458-434b-85ad-98d2253d4d7b-util\") on node \"crc\" DevicePath \"\"" Dec 09 09:57:04 crc kubenswrapper[4824]: I1209 09:57:04.376857 4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d28511a4-2458-434b-85ad-98d2253d4d7b-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 09:57:04 crc kubenswrapper[4824]: I1209 09:57:04.376872 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tptbv\" (UniqueName: \"kubernetes.io/projected/d28511a4-2458-434b-85ad-98d2253d4d7b-kube-api-access-tptbv\") on node \"crc\" DevicePath \"\"" Dec 09 09:57:04 crc kubenswrapper[4824]: I1209 09:57:04.918179 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8" event={"ID":"d28511a4-2458-434b-85ad-98d2253d4d7b","Type":"ContainerDied","Data":"21065b9b332c0d268044dace87b98390c310d6ba7ca63f0aa6cac94bdba87b00"} Dec 09 09:57:04 crc kubenswrapper[4824]: I1209 09:57:04.918243 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21065b9b332c0d268044dace87b98390c310d6ba7ca63f0aa6cac94bdba87b00" Dec 09 09:57:04 crc kubenswrapper[4824]: I1209 09:57:04.918292 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8" Dec 09 09:57:08 crc kubenswrapper[4824]: I1209 09:57:08.957943 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vvxck"] Dec 09 09:57:08 crc kubenswrapper[4824]: I1209 09:57:08.960621 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="nbdb" containerID="cri-o://db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c" gracePeriod=30 Dec 09 09:57:08 crc kubenswrapper[4824]: I1209 09:57:08.960684 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8" gracePeriod=30 Dec 09 09:57:08 crc kubenswrapper[4824]: I1209 09:57:08.960663 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="kube-rbac-proxy-node" containerID="cri-o://8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8" gracePeriod=30 Dec 09 09:57:08 crc kubenswrapper[4824]: I1209 09:57:08.960915 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="northd" containerID="cri-o://02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96" gracePeriod=30 Dec 09 09:57:08 crc kubenswrapper[4824]: I1209 09:57:08.961055 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="ovn-acl-logging" containerID="cri-o://64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f" gracePeriod=30 Dec 09 09:57:08 crc kubenswrapper[4824]: I1209 09:57:08.961136 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="sbdb" containerID="cri-o://c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f" gracePeriod=30 Dec 09 09:57:08 crc kubenswrapper[4824]: I1209 09:57:08.960564 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="ovn-controller" containerID="cri-o://1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695" gracePeriod=30 Dec 09 09:57:09 crc kubenswrapper[4824]: I1209 09:57:09.010918 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="ovnkube-controller" containerID="cri-o://78f6ecf0c19cc7fb61096fcb23d74d6682beca46786235c4a23199197f6757a5" gracePeriod=30 Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:09.956695 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vvxck_13620d58-6925-48c7-bbc7-f34e92922480/ovnkube-controller/3.log" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:09.959404 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vvxck_13620d58-6925-48c7-bbc7-f34e92922480/ovn-acl-logging/0.log" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:09.960225 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vvxck_13620d58-6925-48c7-bbc7-f34e92922480/ovn-controller/0.log" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:09.960885 4824 generic.go:334] "Generic (PLEG): container finished" podID="13620d58-6925-48c7-bbc7-f34e92922480" containerID="78f6ecf0c19cc7fb61096fcb23d74d6682beca46786235c4a23199197f6757a5" exitCode=0 Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:09.960921 4824 generic.go:334] "Generic (PLEG): container finished" podID="13620d58-6925-48c7-bbc7-f34e92922480" containerID="c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f" exitCode=0 Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:09.960931 4824 generic.go:334] "Generic (PLEG): container finished" podID="13620d58-6925-48c7-bbc7-f34e92922480" containerID="db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c" exitCode=0 Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:09.960938 4824 generic.go:334] "Generic (PLEG): container finished" podID="13620d58-6925-48c7-bbc7-f34e92922480" containerID="02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96" exitCode=0 Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:09.960945 4824 generic.go:334] "Generic (PLEG): container finished" podID="13620d58-6925-48c7-bbc7-f34e92922480" containerID="64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f" exitCode=143 Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:09.960953 4824 generic.go:334] "Generic (PLEG): container finished" podID="13620d58-6925-48c7-bbc7-f34e92922480" containerID="1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695" exitCode=143 Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:09.960996 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerDied","Data":"78f6ecf0c19cc7fb61096fcb23d74d6682beca46786235c4a23199197f6757a5"} Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:09.961079 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerDied","Data":"c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f"} Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:09.961100 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerDied","Data":"db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c"} Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:09.961117 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerDied","Data":"02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96"} Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:09.961131 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerDied","Data":"64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f"} Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:09.961140 4824 scope.go:117] "RemoveContainer" containerID="55659b94cb673db2b20bd35a15232443b44882962e7731116b74c418ea5d251e" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:09.961150 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerDied","Data":"1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695"} Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:09.963402 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-pvksg_74aa3cd7-0abc-4fc9-9697-e23b693a7611/kube-multus/2.log" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:09.963859 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-pvksg_74aa3cd7-0abc-4fc9-9697-e23b693a7611/kube-multus/1.log" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:09.963890 4824 generic.go:334] "Generic (PLEG): container finished" podID="74aa3cd7-0abc-4fc9-9697-e23b693a7611" containerID="bf4f9aab454720c3762d743e8301467299a78c44fb18b35f40c349f4d60497d7" exitCode=2 Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:09.963924 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pvksg" event={"ID":"74aa3cd7-0abc-4fc9-9697-e23b693a7611","Type":"ContainerDied","Data":"bf4f9aab454720c3762d743e8301467299a78c44fb18b35f40c349f4d60497d7"} Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:09.964750 4824 scope.go:117] "RemoveContainer" containerID="bf4f9aab454720c3762d743e8301467299a78c44fb18b35f40c349f4d60497d7" Dec 09 09:57:10 crc kubenswrapper[4824]: E1209 09:57:09.965167 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-pvksg_openshift-multus(74aa3cd7-0abc-4fc9-9697-e23b693a7611)\"" pod="openshift-multus/multus-pvksg" podUID="74aa3cd7-0abc-4fc9-9697-e23b693a7611" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.301681 4824 scope.go:117] "RemoveContainer" containerID="c0ee66e89363676597efd13d02245ce8c9a9af288c04f2ed9b7ae8bed31da259" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.807154 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vvxck_13620d58-6925-48c7-bbc7-f34e92922480/ovn-acl-logging/0.log" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.807855 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vvxck_13620d58-6925-48c7-bbc7-f34e92922480/ovn-controller/0.log" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.808298 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894039 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-etc-openvswitch\") pod \"13620d58-6925-48c7-bbc7-f34e92922480\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894111 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-run-ovn\") pod \"13620d58-6925-48c7-bbc7-f34e92922480\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894145 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-run-ovn-kubernetes\") pod \"13620d58-6925-48c7-bbc7-f34e92922480\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894163 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-node-log\") pod \"13620d58-6925-48c7-bbc7-f34e92922480\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894213 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-run-systemd\") pod \"13620d58-6925-48c7-bbc7-f34e92922480\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894247 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/13620d58-6925-48c7-bbc7-f34e92922480-env-overrides\") pod \"13620d58-6925-48c7-bbc7-f34e92922480\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894271 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-systemd-units\") pod \"13620d58-6925-48c7-bbc7-f34e92922480\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894244 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "13620d58-6925-48c7-bbc7-f34e92922480" (UID: "13620d58-6925-48c7-bbc7-f34e92922480"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894301 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-slash\") pod \"13620d58-6925-48c7-bbc7-f34e92922480\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894244 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "13620d58-6925-48c7-bbc7-f34e92922480" (UID: "13620d58-6925-48c7-bbc7-f34e92922480"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894327 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-node-log" (OuterVolumeSpecName: "node-log") pod "13620d58-6925-48c7-bbc7-f34e92922480" (UID: "13620d58-6925-48c7-bbc7-f34e92922480"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894349 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "13620d58-6925-48c7-bbc7-f34e92922480" (UID: "13620d58-6925-48c7-bbc7-f34e92922480"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894366 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-log-socket\") pod \"13620d58-6925-48c7-bbc7-f34e92922480\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894397 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-slash" (OuterVolumeSpecName: "host-slash") pod "13620d58-6925-48c7-bbc7-f34e92922480" (UID: "13620d58-6925-48c7-bbc7-f34e92922480"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894384 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "13620d58-6925-48c7-bbc7-f34e92922480" (UID: "13620d58-6925-48c7-bbc7-f34e92922480"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894448 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-log-socket" (OuterVolumeSpecName: "log-socket") pod "13620d58-6925-48c7-bbc7-f34e92922480" (UID: "13620d58-6925-48c7-bbc7-f34e92922480"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894454 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "13620d58-6925-48c7-bbc7-f34e92922480" (UID: "13620d58-6925-48c7-bbc7-f34e92922480"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894422 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-run-openvswitch\") pod \"13620d58-6925-48c7-bbc7-f34e92922480\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894500 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-kubelet\") pod \"13620d58-6925-48c7-bbc7-f34e92922480\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894517 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "13620d58-6925-48c7-bbc7-f34e92922480" (UID: "13620d58-6925-48c7-bbc7-f34e92922480"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894540 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrd4m\" (UniqueName: \"kubernetes.io/projected/13620d58-6925-48c7-bbc7-f34e92922480-kube-api-access-zrd4m\") pod \"13620d58-6925-48c7-bbc7-f34e92922480\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894580 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/13620d58-6925-48c7-bbc7-f34e92922480-ovn-node-metrics-cert\") pod \"13620d58-6925-48c7-bbc7-f34e92922480\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894613 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-cni-bin\") pod \"13620d58-6925-48c7-bbc7-f34e92922480\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894694 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-var-lib-cni-networks-ovn-kubernetes\") pod \"13620d58-6925-48c7-bbc7-f34e92922480\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894738 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-run-netns\") pod \"13620d58-6925-48c7-bbc7-f34e92922480\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894835 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-var-lib-openvswitch\") pod \"13620d58-6925-48c7-bbc7-f34e92922480\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894837 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "13620d58-6925-48c7-bbc7-f34e92922480" (UID: "13620d58-6925-48c7-bbc7-f34e92922480"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894842 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "13620d58-6925-48c7-bbc7-f34e92922480" (UID: "13620d58-6925-48c7-bbc7-f34e92922480"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894883 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/13620d58-6925-48c7-bbc7-f34e92922480-ovnkube-script-lib\") pod \"13620d58-6925-48c7-bbc7-f34e92922480\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894875 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "13620d58-6925-48c7-bbc7-f34e92922480" (UID: "13620d58-6925-48c7-bbc7-f34e92922480"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894958 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/13620d58-6925-48c7-bbc7-f34e92922480-ovnkube-config\") pod \"13620d58-6925-48c7-bbc7-f34e92922480\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.894999 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-cni-netd\") pod \"13620d58-6925-48c7-bbc7-f34e92922480\" (UID: \"13620d58-6925-48c7-bbc7-f34e92922480\") " Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.895066 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13620d58-6925-48c7-bbc7-f34e92922480-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "13620d58-6925-48c7-bbc7-f34e92922480" (UID: "13620d58-6925-48c7-bbc7-f34e92922480"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.895180 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "13620d58-6925-48c7-bbc7-f34e92922480" (UID: "13620d58-6925-48c7-bbc7-f34e92922480"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.895311 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13620d58-6925-48c7-bbc7-f34e92922480-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "13620d58-6925-48c7-bbc7-f34e92922480" (UID: "13620d58-6925-48c7-bbc7-f34e92922480"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.895379 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "13620d58-6925-48c7-bbc7-f34e92922480" (UID: "13620d58-6925-48c7-bbc7-f34e92922480"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.895416 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13620d58-6925-48c7-bbc7-f34e92922480-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "13620d58-6925-48c7-bbc7-f34e92922480" (UID: "13620d58-6925-48c7-bbc7-f34e92922480"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.895719 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/13620d58-6925-48c7-bbc7-f34e92922480-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.895740 4824 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.895752 4824 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.895811 4824 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.895830 4824 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.895846 4824 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-node-log\") on node \"crc\" DevicePath \"\"" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.895858 4824 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/13620d58-6925-48c7-bbc7-f34e92922480-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.895869 4824 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.895880 4824 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-slash\") on node \"crc\" DevicePath \"\"" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.895892 4824 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-log-socket\") on node \"crc\" DevicePath \"\"" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.895904 4824 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.895915 4824 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.895928 4824 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.895943 4824 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.895956 4824 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.895969 4824 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.895980 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/13620d58-6925-48c7-bbc7-f34e92922480-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.898742 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2qqhd"] Dec 09 09:57:10 crc kubenswrapper[4824]: E1209 09:57:10.902338 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="sbdb" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902367 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="sbdb" Dec 09 09:57:10 crc kubenswrapper[4824]: E1209 09:57:10.902380 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="ovnkube-controller" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902390 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="ovnkube-controller" Dec 09 09:57:10 crc kubenswrapper[4824]: E1209 09:57:10.902402 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="ovnkube-controller" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902410 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="ovnkube-controller" Dec 09 09:57:10 crc kubenswrapper[4824]: E1209 09:57:10.902424 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d28511a4-2458-434b-85ad-98d2253d4d7b" containerName="util" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902431 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d28511a4-2458-434b-85ad-98d2253d4d7b" containerName="util" Dec 09 09:57:10 crc kubenswrapper[4824]: E1209 09:57:10.902441 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="northd" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902448 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="northd" Dec 09 09:57:10 crc kubenswrapper[4824]: E1209 09:57:10.902457 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="nbdb" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902462 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="nbdb" Dec 09 09:57:10 crc kubenswrapper[4824]: E1209 09:57:10.902473 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d28511a4-2458-434b-85ad-98d2253d4d7b" containerName="extract" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902479 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d28511a4-2458-434b-85ad-98d2253d4d7b" containerName="extract" Dec 09 09:57:10 crc kubenswrapper[4824]: E1209 09:57:10.902488 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="ovn-acl-logging" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902495 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="ovn-acl-logging" Dec 09 09:57:10 crc kubenswrapper[4824]: E1209 09:57:10.902504 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d28511a4-2458-434b-85ad-98d2253d4d7b" containerName="pull" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902510 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d28511a4-2458-434b-85ad-98d2253d4d7b" containerName="pull" Dec 09 09:57:10 crc kubenswrapper[4824]: E1209 09:57:10.902520 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="kube-rbac-proxy-ovn-metrics" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902527 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="kube-rbac-proxy-ovn-metrics" Dec 09 09:57:10 crc kubenswrapper[4824]: E1209 09:57:10.902537 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="kube-rbac-proxy-node" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902551 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="kube-rbac-proxy-node" Dec 09 09:57:10 crc kubenswrapper[4824]: E1209 09:57:10.902568 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="kubecfg-setup" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902576 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="kubecfg-setup" Dec 09 09:57:10 crc kubenswrapper[4824]: E1209 09:57:10.902595 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="ovn-controller" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902601 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="ovn-controller" Dec 09 09:57:10 crc kubenswrapper[4824]: E1209 09:57:10.902608 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="ovnkube-controller" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902614 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="ovnkube-controller" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902754 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="ovnkube-controller" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902769 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="kube-rbac-proxy-node" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902783 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d28511a4-2458-434b-85ad-98d2253d4d7b" containerName="extract" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902807 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="kube-rbac-proxy-ovn-metrics" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902816 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="sbdb" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902824 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="ovnkube-controller" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902834 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="ovnkube-controller" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902843 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="northd" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902851 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="ovn-acl-logging" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902860 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="nbdb" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902868 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="ovn-controller" Dec 09 09:57:10 crc kubenswrapper[4824]: E1209 09:57:10.902979 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="ovnkube-controller" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.902987 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="ovnkube-controller" Dec 09 09:57:10 crc kubenswrapper[4824]: E1209 09:57:10.902998 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="ovnkube-controller" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.903004 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="ovnkube-controller" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.903131 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="ovnkube-controller" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.903143 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="13620d58-6925-48c7-bbc7-f34e92922480" containerName="ovnkube-controller" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.905296 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.907538 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13620d58-6925-48c7-bbc7-f34e92922480-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "13620d58-6925-48c7-bbc7-f34e92922480" (UID: "13620d58-6925-48c7-bbc7-f34e92922480"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.917146 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13620d58-6925-48c7-bbc7-f34e92922480-kube-api-access-zrd4m" (OuterVolumeSpecName: "kube-api-access-zrd4m") pod "13620d58-6925-48c7-bbc7-f34e92922480" (UID: "13620d58-6925-48c7-bbc7-f34e92922480"). InnerVolumeSpecName "kube-api-access-zrd4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.925431 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "13620d58-6925-48c7-bbc7-f34e92922480" (UID: "13620d58-6925-48c7-bbc7-f34e92922480"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.972854 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-pvksg_74aa3cd7-0abc-4fc9-9697-e23b693a7611/kube-multus/2.log" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.978728 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vvxck_13620d58-6925-48c7-bbc7-f34e92922480/ovn-acl-logging/0.log" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.979445 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vvxck_13620d58-6925-48c7-bbc7-f34e92922480/ovn-controller/0.log" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.979941 4824 generic.go:334] "Generic (PLEG): container finished" podID="13620d58-6925-48c7-bbc7-f34e92922480" containerID="d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8" exitCode=0 Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.979973 4824 generic.go:334] "Generic (PLEG): container finished" podID="13620d58-6925-48c7-bbc7-f34e92922480" containerID="8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8" exitCode=0 Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.980008 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerDied","Data":"d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8"} Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.980045 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerDied","Data":"8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8"} Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.980055 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" event={"ID":"13620d58-6925-48c7-bbc7-f34e92922480","Type":"ContainerDied","Data":"b0d26dac1a3a6325cac749ee337191fb4c11cb083c7dfc4a4afc10c4310fdbb3"} Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.980075 4824 scope.go:117] "RemoveContainer" containerID="78f6ecf0c19cc7fb61096fcb23d74d6682beca46786235c4a23199197f6757a5" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.980095 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vvxck" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.997299 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-var-lib-openvswitch\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.997341 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-run-systemd\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.997363 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-log-socket\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.997426 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-run-ovn\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.997525 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-host-cni-netd\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.997558 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-systemd-units\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.997589 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9f7b\" (UniqueName: \"kubernetes.io/projected/34793d58-1049-48ff-b8e7-187bdf42e550-kube-api-access-f9f7b\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.997631 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-run-openvswitch\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.997665 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-host-run-ovn-kubernetes\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.997890 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-node-log\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.997963 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.998019 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-etc-openvswitch\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.998225 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-host-kubelet\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.998294 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/34793d58-1049-48ff-b8e7-187bdf42e550-env-overrides\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.998338 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/34793d58-1049-48ff-b8e7-187bdf42e550-ovnkube-config\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.998405 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-host-run-netns\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.998446 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-host-cni-bin\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.998467 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/34793d58-1049-48ff-b8e7-187bdf42e550-ovnkube-script-lib\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.998489 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-host-slash\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.998508 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/34793d58-1049-48ff-b8e7-187bdf42e550-ovn-node-metrics-cert\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.998737 4824 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/13620d58-6925-48c7-bbc7-f34e92922480-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.998771 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrd4m\" (UniqueName: \"kubernetes.io/projected/13620d58-6925-48c7-bbc7-f34e92922480-kube-api-access-zrd4m\") on node \"crc\" DevicePath \"\"" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.998807 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/13620d58-6925-48c7-bbc7-f34e92922480-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 09 09:57:10 crc kubenswrapper[4824]: I1209 09:57:10.999566 4824 scope.go:117] "RemoveContainer" containerID="c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.022115 4824 scope.go:117] "RemoveContainer" containerID="db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.043142 4824 scope.go:117] "RemoveContainer" containerID="02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.072510 4824 scope.go:117] "RemoveContainer" containerID="d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.073950 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vvxck"] Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.084665 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vvxck"] Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.102348 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-var-lib-openvswitch\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.102414 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-run-systemd\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.102443 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-log-socket\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.102470 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-run-ovn\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.102504 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-host-cni-netd\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.102526 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-systemd-units\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.102552 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9f7b\" (UniqueName: \"kubernetes.io/projected/34793d58-1049-48ff-b8e7-187bdf42e550-kube-api-access-f9f7b\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.102592 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-run-openvswitch\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.102614 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-host-run-ovn-kubernetes\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.102646 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-node-log\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.102668 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.102692 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-etc-openvswitch\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.102722 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-host-kubelet\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.102752 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/34793d58-1049-48ff-b8e7-187bdf42e550-env-overrides\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.102772 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/34793d58-1049-48ff-b8e7-187bdf42e550-ovnkube-config\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.102821 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-host-run-netns\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.102857 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-host-cni-bin\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.102903 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/34793d58-1049-48ff-b8e7-187bdf42e550-ovnkube-script-lib\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.102932 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-host-slash\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.102952 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/34793d58-1049-48ff-b8e7-187bdf42e550-ovn-node-metrics-cert\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.104967 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-node-log\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.105085 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-var-lib-openvswitch\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.105147 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-run-systemd\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.105188 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-log-socket\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.105215 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-run-ovn\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.105244 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-host-cni-netd\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.105268 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-systemd-units\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.105662 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-run-openvswitch\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.105699 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-host-run-ovn-kubernetes\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.105731 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-host-cni-bin\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.105759 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-host-run-netns\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.105804 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-etc-openvswitch\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.105829 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.105856 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-host-kubelet\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.105912 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/34793d58-1049-48ff-b8e7-187bdf42e550-ovnkube-config\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.106440 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/34793d58-1049-48ff-b8e7-187bdf42e550-env-overrides\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.106507 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/34793d58-1049-48ff-b8e7-187bdf42e550-host-slash\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.106598 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/34793d58-1049-48ff-b8e7-187bdf42e550-ovnkube-script-lib\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.113603 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/34793d58-1049-48ff-b8e7-187bdf42e550-ovn-node-metrics-cert\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.122096 4824 scope.go:117] "RemoveContainer" containerID="8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.139569 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9f7b\" (UniqueName: \"kubernetes.io/projected/34793d58-1049-48ff-b8e7-187bdf42e550-kube-api-access-f9f7b\") pod \"ovnkube-node-2qqhd\" (UID: \"34793d58-1049-48ff-b8e7-187bdf42e550\") " pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.181934 4824 scope.go:117] "RemoveContainer" containerID="64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.203599 4824 scope.go:117] "RemoveContainer" containerID="1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.227038 4824 scope.go:117] "RemoveContainer" containerID="21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.241728 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.299230 4824 scope.go:117] "RemoveContainer" containerID="78f6ecf0c19cc7fb61096fcb23d74d6682beca46786235c4a23199197f6757a5" Dec 09 09:57:11 crc kubenswrapper[4824]: E1209 09:57:11.299798 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78f6ecf0c19cc7fb61096fcb23d74d6682beca46786235c4a23199197f6757a5\": container with ID starting with 78f6ecf0c19cc7fb61096fcb23d74d6682beca46786235c4a23199197f6757a5 not found: ID does not exist" containerID="78f6ecf0c19cc7fb61096fcb23d74d6682beca46786235c4a23199197f6757a5" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.299834 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78f6ecf0c19cc7fb61096fcb23d74d6682beca46786235c4a23199197f6757a5"} err="failed to get container status \"78f6ecf0c19cc7fb61096fcb23d74d6682beca46786235c4a23199197f6757a5\": rpc error: code = NotFound desc = could not find container \"78f6ecf0c19cc7fb61096fcb23d74d6682beca46786235c4a23199197f6757a5\": container with ID starting with 78f6ecf0c19cc7fb61096fcb23d74d6682beca46786235c4a23199197f6757a5 not found: ID does not exist" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.299868 4824 scope.go:117] "RemoveContainer" containerID="c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f" Dec 09 09:57:11 crc kubenswrapper[4824]: E1209 09:57:11.300456 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\": container with ID starting with c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f not found: ID does not exist" containerID="c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.300488 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f"} err="failed to get container status \"c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\": rpc error: code = NotFound desc = could not find container \"c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\": container with ID starting with c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f not found: ID does not exist" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.300508 4824 scope.go:117] "RemoveContainer" containerID="db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c" Dec 09 09:57:11 crc kubenswrapper[4824]: E1209 09:57:11.300953 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\": container with ID starting with db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c not found: ID does not exist" containerID="db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.300982 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c"} err="failed to get container status \"db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\": rpc error: code = NotFound desc = could not find container \"db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\": container with ID starting with db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c not found: ID does not exist" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.301008 4824 scope.go:117] "RemoveContainer" containerID="02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96" Dec 09 09:57:11 crc kubenswrapper[4824]: E1209 09:57:11.302212 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\": container with ID starting with 02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96 not found: ID does not exist" containerID="02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.302236 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96"} err="failed to get container status \"02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\": rpc error: code = NotFound desc = could not find container \"02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\": container with ID starting with 02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96 not found: ID does not exist" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.302251 4824 scope.go:117] "RemoveContainer" containerID="d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8" Dec 09 09:57:11 crc kubenswrapper[4824]: E1209 09:57:11.302583 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\": container with ID starting with d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8 not found: ID does not exist" containerID="d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.302606 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8"} err="failed to get container status \"d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\": rpc error: code = NotFound desc = could not find container \"d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\": container with ID starting with d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8 not found: ID does not exist" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.302620 4824 scope.go:117] "RemoveContainer" containerID="8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8" Dec 09 09:57:11 crc kubenswrapper[4824]: E1209 09:57:11.302941 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\": container with ID starting with 8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8 not found: ID does not exist" containerID="8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.302966 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8"} err="failed to get container status \"8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\": rpc error: code = NotFound desc = could not find container \"8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\": container with ID starting with 8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8 not found: ID does not exist" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.302985 4824 scope.go:117] "RemoveContainer" containerID="64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f" Dec 09 09:57:11 crc kubenswrapper[4824]: E1209 09:57:11.303326 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\": container with ID starting with 64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f not found: ID does not exist" containerID="64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.303350 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f"} err="failed to get container status \"64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\": rpc error: code = NotFound desc = could not find container \"64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\": container with ID starting with 64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f not found: ID does not exist" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.303365 4824 scope.go:117] "RemoveContainer" containerID="1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695" Dec 09 09:57:11 crc kubenswrapper[4824]: E1209 09:57:11.303831 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\": container with ID starting with 1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695 not found: ID does not exist" containerID="1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.303887 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695"} err="failed to get container status \"1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\": rpc error: code = NotFound desc = could not find container \"1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\": container with ID starting with 1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695 not found: ID does not exist" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.303929 4824 scope.go:117] "RemoveContainer" containerID="21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339" Dec 09 09:57:11 crc kubenswrapper[4824]: E1209 09:57:11.304307 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\": container with ID starting with 21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339 not found: ID does not exist" containerID="21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.304333 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339"} err="failed to get container status \"21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\": rpc error: code = NotFound desc = could not find container \"21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\": container with ID starting with 21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339 not found: ID does not exist" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.304352 4824 scope.go:117] "RemoveContainer" containerID="78f6ecf0c19cc7fb61096fcb23d74d6682beca46786235c4a23199197f6757a5" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.304704 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78f6ecf0c19cc7fb61096fcb23d74d6682beca46786235c4a23199197f6757a5"} err="failed to get container status \"78f6ecf0c19cc7fb61096fcb23d74d6682beca46786235c4a23199197f6757a5\": rpc error: code = NotFound desc = could not find container \"78f6ecf0c19cc7fb61096fcb23d74d6682beca46786235c4a23199197f6757a5\": container with ID starting with 78f6ecf0c19cc7fb61096fcb23d74d6682beca46786235c4a23199197f6757a5 not found: ID does not exist" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.304768 4824 scope.go:117] "RemoveContainer" containerID="c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.305304 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f"} err="failed to get container status \"c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\": rpc error: code = NotFound desc = could not find container \"c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f\": container with ID starting with c5b6990572ef3aafc2a710b6a8984311177d2411bfa348e63c8c9d4a4dc39b8f not found: ID does not exist" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.305400 4824 scope.go:117] "RemoveContainer" containerID="db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.306177 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c"} err="failed to get container status \"db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\": rpc error: code = NotFound desc = could not find container \"db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c\": container with ID starting with db4c50f9c278e66381741484571b4eef9519c2f6977b258a43aa1c97149bfe4c not found: ID does not exist" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.306216 4824 scope.go:117] "RemoveContainer" containerID="02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.308496 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96"} err="failed to get container status \"02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\": rpc error: code = NotFound desc = could not find container \"02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96\": container with ID starting with 02a1145d4e188129e7b995c47a8995a48baa9b8b13e134a592325cef064e0e96 not found: ID does not exist" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.308531 4824 scope.go:117] "RemoveContainer" containerID="d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.308865 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8"} err="failed to get container status \"d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\": rpc error: code = NotFound desc = could not find container \"d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8\": container with ID starting with d6e3e857f571a0ad1fd17d3c313e455e80d3412dea29b3e943e0b6ffb0a6e6a8 not found: ID does not exist" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.308914 4824 scope.go:117] "RemoveContainer" containerID="8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.309202 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8"} err="failed to get container status \"8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\": rpc error: code = NotFound desc = could not find container \"8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8\": container with ID starting with 8503e081337afefc28bfaa85f99f18ee39513b400ca1fcdf78bbadfcfa8278b8 not found: ID does not exist" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.309235 4824 scope.go:117] "RemoveContainer" containerID="64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.309530 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f"} err="failed to get container status \"64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\": rpc error: code = NotFound desc = could not find container \"64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f\": container with ID starting with 64ac70085df90ff833a7056d4afa4a3c043401553e848154ae515e9247d63e8f not found: ID does not exist" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.309554 4824 scope.go:117] "RemoveContainer" containerID="1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.309825 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695"} err="failed to get container status \"1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\": rpc error: code = NotFound desc = could not find container \"1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695\": container with ID starting with 1c48a3c10f82c35fefd956cc6851a4cfba3dcacb2f3895e1428279edf329b695 not found: ID does not exist" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.309844 4824 scope.go:117] "RemoveContainer" containerID="21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.310100 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339"} err="failed to get container status \"21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\": rpc error: code = NotFound desc = could not find container \"21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339\": container with ID starting with 21ca901e6c36057674ce0a594d7bbf2293620cee578727f3a8780a7834b43339 not found: ID does not exist" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.919731 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13620d58-6925-48c7-bbc7-f34e92922480" path="/var/lib/kubelet/pods/13620d58-6925-48c7-bbc7-f34e92922480/volumes" Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.994086 4824 generic.go:334] "Generic (PLEG): container finished" podID="34793d58-1049-48ff-b8e7-187bdf42e550" containerID="168b62963865115275fde4ba456c232e1e7f655ddc88f52edefacb31ce0753ac" exitCode=0 Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.994169 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" event={"ID":"34793d58-1049-48ff-b8e7-187bdf42e550","Type":"ContainerDied","Data":"168b62963865115275fde4ba456c232e1e7f655ddc88f52edefacb31ce0753ac"} Dec 09 09:57:11 crc kubenswrapper[4824]: I1209 09:57:11.994250 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" event={"ID":"34793d58-1049-48ff-b8e7-187bdf42e550","Type":"ContainerStarted","Data":"0e181aa056e5d068566c698c905e652e6df558d55cc1770160b37716e26d7be7"} Dec 09 09:57:13 crc kubenswrapper[4824]: I1209 09:57:13.007184 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" event={"ID":"34793d58-1049-48ff-b8e7-187bdf42e550","Type":"ContainerStarted","Data":"2fb42c39c0b7c4694b78525af297d22be0f5e8dc03430a2cb78bff291bd74466"} Dec 09 09:57:13 crc kubenswrapper[4824]: I1209 09:57:13.007631 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" event={"ID":"34793d58-1049-48ff-b8e7-187bdf42e550","Type":"ContainerStarted","Data":"04793b7c8ad4cc584b95e648d409a193be029a51c57570c09ff8ca8b545b8512"} Dec 09 09:57:13 crc kubenswrapper[4824]: I1209 09:57:13.007644 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" event={"ID":"34793d58-1049-48ff-b8e7-187bdf42e550","Type":"ContainerStarted","Data":"e8e62f7b6012596ed9f2b8be0b5c014586397dc5e787124a0bc6eb8d592b4bcc"} Dec 09 09:57:14 crc kubenswrapper[4824]: I1209 09:57:14.042758 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" event={"ID":"34793d58-1049-48ff-b8e7-187bdf42e550","Type":"ContainerStarted","Data":"b0519123b56c8d15752c4a8adfcf2d8d2234222706cec2073884e903bdb96710"} Dec 09 09:57:15 crc kubenswrapper[4824]: I1209 09:57:15.055439 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" event={"ID":"34793d58-1049-48ff-b8e7-187bdf42e550","Type":"ContainerStarted","Data":"e62e1d724dfb76dd9cec2788df08613039c2d195701893ad8c7d0a3a5b5337df"} Dec 09 09:57:15 crc kubenswrapper[4824]: I1209 09:57:15.055912 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" event={"ID":"34793d58-1049-48ff-b8e7-187bdf42e550","Type":"ContainerStarted","Data":"264e3e1c6a29abcb1229990952978bee166a8a838f17d8c8c581c2066b85739b"} Dec 09 09:57:18 crc kubenswrapper[4824]: I1209 09:57:18.117539 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" event={"ID":"34793d58-1049-48ff-b8e7-187bdf42e550","Type":"ContainerStarted","Data":"bff22bf190a57dc10e0cf167e2bb9b0f26338352a3de07510063a1985857b088"} Dec 09 09:57:18 crc kubenswrapper[4824]: I1209 09:57:18.799595 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-xb2z8"] Dec 09 09:57:18 crc kubenswrapper[4824]: I1209 09:57:18.801301 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-xb2z8" Dec 09 09:57:18 crc kubenswrapper[4824]: I1209 09:57:18.805964 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 09 09:57:18 crc kubenswrapper[4824]: I1209 09:57:18.806845 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 09 09:57:18 crc kubenswrapper[4824]: I1209 09:57:18.807069 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-spgh2" Dec 09 09:57:18 crc kubenswrapper[4824]: I1209 09:57:18.899804 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp"] Dec 09 09:57:18 crc kubenswrapper[4824]: I1209 09:57:18.901113 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" Dec 09 09:57:18 crc kubenswrapper[4824]: I1209 09:57:18.906848 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf"] Dec 09 09:57:18 crc kubenswrapper[4824]: I1209 09:57:18.908053 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf" Dec 09 09:57:18 crc kubenswrapper[4824]: I1209 09:57:18.908543 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pxbj\" (UniqueName: \"kubernetes.io/projected/ca1c7fd5-ccf0-453d-9426-d77e591d729e-kube-api-access-6pxbj\") pod \"obo-prometheus-operator-668cf9dfbb-xb2z8\" (UID: \"ca1c7fd5-ccf0-453d-9426-d77e591d729e\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-xb2z8" Dec 09 09:57:18 crc kubenswrapper[4824]: I1209 09:57:18.908906 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-c2pj6" Dec 09 09:57:18 crc kubenswrapper[4824]: I1209 09:57:18.912343 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.011025 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pxbj\" (UniqueName: \"kubernetes.io/projected/ca1c7fd5-ccf0-453d-9426-d77e591d729e-kube-api-access-6pxbj\") pod \"obo-prometheus-operator-668cf9dfbb-xb2z8\" (UID: \"ca1c7fd5-ccf0-453d-9426-d77e591d729e\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-xb2z8" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.011101 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cdadfe90-b4b1-4c57-a187-f6e2cbb72103-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp\" (UID: \"cdadfe90-b4b1-4c57-a187-f6e2cbb72103\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.011136 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cdadfe90-b4b1-4c57-a187-f6e2cbb72103-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp\" (UID: \"cdadfe90-b4b1-4c57-a187-f6e2cbb72103\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.011241 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/eb813060-c1d8-433f-be2e-e85d102499d9-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf\" (UID: \"eb813060-c1d8-433f-be2e-e85d102499d9\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.011683 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eb813060-c1d8-433f-be2e-e85d102499d9-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf\" (UID: \"eb813060-c1d8-433f-be2e-e85d102499d9\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.056897 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-qn2qz"] Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.058305 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.062778 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.063715 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-s2bhp" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.130754 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eb813060-c1d8-433f-be2e-e85d102499d9-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf\" (UID: \"eb813060-c1d8-433f-be2e-e85d102499d9\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.130896 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cdadfe90-b4b1-4c57-a187-f6e2cbb72103-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp\" (UID: \"cdadfe90-b4b1-4c57-a187-f6e2cbb72103\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.130962 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cdadfe90-b4b1-4c57-a187-f6e2cbb72103-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp\" (UID: \"cdadfe90-b4b1-4c57-a187-f6e2cbb72103\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.131009 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/eb813060-c1d8-433f-be2e-e85d102499d9-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf\" (UID: \"eb813060-c1d8-433f-be2e-e85d102499d9\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.136668 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/eb813060-c1d8-433f-be2e-e85d102499d9-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf\" (UID: \"eb813060-c1d8-433f-be2e-e85d102499d9\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.138483 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pxbj\" (UniqueName: \"kubernetes.io/projected/ca1c7fd5-ccf0-453d-9426-d77e591d729e-kube-api-access-6pxbj\") pod \"obo-prometheus-operator-668cf9dfbb-xb2z8\" (UID: \"ca1c7fd5-ccf0-453d-9426-d77e591d729e\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-xb2z8" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.140704 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eb813060-c1d8-433f-be2e-e85d102499d9-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf\" (UID: \"eb813060-c1d8-433f-be2e-e85d102499d9\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.142763 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cdadfe90-b4b1-4c57-a187-f6e2cbb72103-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp\" (UID: \"cdadfe90-b4b1-4c57-a187-f6e2cbb72103\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.143083 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-xb2z8" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.143911 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cdadfe90-b4b1-4c57-a187-f6e2cbb72103-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp\" (UID: \"cdadfe90-b4b1-4c57-a187-f6e2cbb72103\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.253249 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6n55\" (UniqueName: \"kubernetes.io/projected/612ebf47-a6c4-4a62-8d97-91a003f49c44-kube-api-access-r6n55\") pod \"observability-operator-d8bb48f5d-qn2qz\" (UID: \"612ebf47-a6c4-4a62-8d97-91a003f49c44\") " pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.253393 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/612ebf47-a6c4-4a62-8d97-91a003f49c44-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-qn2qz\" (UID: \"612ebf47-a6c4-4a62-8d97-91a003f49c44\") " pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.253668 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.254474 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" Dec 09 09:57:19 crc kubenswrapper[4824]: E1209 09:57:19.323540 4824 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-xb2z8_openshift-operators_ca1c7fd5-ccf0-453d-9426-d77e591d729e_0(01240aef49d896c16c0f1f6c7612cc9c5faf06e690e0a02f5e14a03c8ba57f31): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 09:57:19 crc kubenswrapper[4824]: E1209 09:57:19.323635 4824 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-xb2z8_openshift-operators_ca1c7fd5-ccf0-453d-9426-d77e591d729e_0(01240aef49d896c16c0f1f6c7612cc9c5faf06e690e0a02f5e14a03c8ba57f31): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-xb2z8" Dec 09 09:57:19 crc kubenswrapper[4824]: E1209 09:57:19.323664 4824 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-xb2z8_openshift-operators_ca1c7fd5-ccf0-453d-9426-d77e591d729e_0(01240aef49d896c16c0f1f6c7612cc9c5faf06e690e0a02f5e14a03c8ba57f31): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-xb2z8" Dec 09 09:57:19 crc kubenswrapper[4824]: E1209 09:57:19.323733 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-xb2z8_openshift-operators(ca1c7fd5-ccf0-453d-9426-d77e591d729e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-xb2z8_openshift-operators(ca1c7fd5-ccf0-453d-9426-d77e591d729e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-xb2z8_openshift-operators_ca1c7fd5-ccf0-453d-9426-d77e591d729e_0(01240aef49d896c16c0f1f6c7612cc9c5faf06e690e0a02f5e14a03c8ba57f31): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-xb2z8" podUID="ca1c7fd5-ccf0-453d-9426-d77e591d729e" Dec 09 09:57:19 crc kubenswrapper[4824]: E1209 09:57:19.333096 4824 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf_openshift-operators_eb813060-c1d8-433f-be2e-e85d102499d9_0(c56fa5d4ad902359c01cfed1cb69adf5cd8a0d5957bf5052a8161a8409adcd15): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 09:57:19 crc kubenswrapper[4824]: E1209 09:57:19.333213 4824 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf_openshift-operators_eb813060-c1d8-433f-be2e-e85d102499d9_0(c56fa5d4ad902359c01cfed1cb69adf5cd8a0d5957bf5052a8161a8409adcd15): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf" Dec 09 09:57:19 crc kubenswrapper[4824]: E1209 09:57:19.333253 4824 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf_openshift-operators_eb813060-c1d8-433f-be2e-e85d102499d9_0(c56fa5d4ad902359c01cfed1cb69adf5cd8a0d5957bf5052a8161a8409adcd15): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf" Dec 09 09:57:19 crc kubenswrapper[4824]: E1209 09:57:19.333345 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf_openshift-operators(eb813060-c1d8-433f-be2e-e85d102499d9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf_openshift-operators(eb813060-c1d8-433f-be2e-e85d102499d9)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf_openshift-operators_eb813060-c1d8-433f-be2e-e85d102499d9_0(c56fa5d4ad902359c01cfed1cb69adf5cd8a0d5957bf5052a8161a8409adcd15): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf" podUID="eb813060-c1d8-433f-be2e-e85d102499d9" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.354555 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/612ebf47-a6c4-4a62-8d97-91a003f49c44-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-qn2qz\" (UID: \"612ebf47-a6c4-4a62-8d97-91a003f49c44\") " pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.354655 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6n55\" (UniqueName: \"kubernetes.io/projected/612ebf47-a6c4-4a62-8d97-91a003f49c44-kube-api-access-r6n55\") pod \"observability-operator-d8bb48f5d-qn2qz\" (UID: \"612ebf47-a6c4-4a62-8d97-91a003f49c44\") " pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.360973 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/612ebf47-a6c4-4a62-8d97-91a003f49c44-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-qn2qz\" (UID: \"612ebf47-a6c4-4a62-8d97-91a003f49c44\") " pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 09:57:19 crc kubenswrapper[4824]: E1209 09:57:19.369211 4824 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp_openshift-operators_cdadfe90-b4b1-4c57-a187-f6e2cbb72103_0(526b97fe6fc1117710127f2d06d7a23ec44c3727b3eeaea87f3ae9985ec11255): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 09:57:19 crc kubenswrapper[4824]: E1209 09:57:19.369326 4824 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp_openshift-operators_cdadfe90-b4b1-4c57-a187-f6e2cbb72103_0(526b97fe6fc1117710127f2d06d7a23ec44c3727b3eeaea87f3ae9985ec11255): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" Dec 09 09:57:19 crc kubenswrapper[4824]: E1209 09:57:19.369365 4824 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp_openshift-operators_cdadfe90-b4b1-4c57-a187-f6e2cbb72103_0(526b97fe6fc1117710127f2d06d7a23ec44c3727b3eeaea87f3ae9985ec11255): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" Dec 09 09:57:19 crc kubenswrapper[4824]: E1209 09:57:19.369441 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp_openshift-operators(cdadfe90-b4b1-4c57-a187-f6e2cbb72103)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp_openshift-operators(cdadfe90-b4b1-4c57-a187-f6e2cbb72103)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp_openshift-operators_cdadfe90-b4b1-4c57-a187-f6e2cbb72103_0(526b97fe6fc1117710127f2d06d7a23ec44c3727b3eeaea87f3ae9985ec11255): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" podUID="cdadfe90-b4b1-4c57-a187-f6e2cbb72103" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.385377 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6n55\" (UniqueName: \"kubernetes.io/projected/612ebf47-a6c4-4a62-8d97-91a003f49c44-kube-api-access-r6n55\") pod \"observability-operator-d8bb48f5d-qn2qz\" (UID: \"612ebf47-a6c4-4a62-8d97-91a003f49c44\") " pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.385945 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.391584 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-pkmss"] Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.393176 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-pkmss" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.396460 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-zxcvt" Dec 09 09:57:19 crc kubenswrapper[4824]: E1209 09:57:19.442163 4824 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-qn2qz_openshift-operators_612ebf47-a6c4-4a62-8d97-91a003f49c44_0(b99df75b867d17ce646bdc0e663d0fc93f26f340fd0105f50dc53e1cb5363a05): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 09:57:19 crc kubenswrapper[4824]: E1209 09:57:19.442260 4824 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-qn2qz_openshift-operators_612ebf47-a6c4-4a62-8d97-91a003f49c44_0(b99df75b867d17ce646bdc0e663d0fc93f26f340fd0105f50dc53e1cb5363a05): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 09:57:19 crc kubenswrapper[4824]: E1209 09:57:19.442309 4824 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-qn2qz_openshift-operators_612ebf47-a6c4-4a62-8d97-91a003f49c44_0(b99df75b867d17ce646bdc0e663d0fc93f26f340fd0105f50dc53e1cb5363a05): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 09:57:19 crc kubenswrapper[4824]: E1209 09:57:19.442377 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-qn2qz_openshift-operators(612ebf47-a6c4-4a62-8d97-91a003f49c44)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-qn2qz_openshift-operators(612ebf47-a6c4-4a62-8d97-91a003f49c44)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-qn2qz_openshift-operators_612ebf47-a6c4-4a62-8d97-91a003f49c44_0(b99df75b867d17ce646bdc0e663d0fc93f26f340fd0105f50dc53e1cb5363a05): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" podUID="612ebf47-a6c4-4a62-8d97-91a003f49c44" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.558767 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/2ecf9e90-37fe-416d-9e7a-400373cfbc8d-openshift-service-ca\") pod \"perses-operator-5446b9c989-pkmss\" (UID: \"2ecf9e90-37fe-416d-9e7a-400373cfbc8d\") " pod="openshift-operators/perses-operator-5446b9c989-pkmss" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.559267 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q55v9\" (UniqueName: \"kubernetes.io/projected/2ecf9e90-37fe-416d-9e7a-400373cfbc8d-kube-api-access-q55v9\") pod \"perses-operator-5446b9c989-pkmss\" (UID: \"2ecf9e90-37fe-416d-9e7a-400373cfbc8d\") " pod="openshift-operators/perses-operator-5446b9c989-pkmss" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.660714 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/2ecf9e90-37fe-416d-9e7a-400373cfbc8d-openshift-service-ca\") pod \"perses-operator-5446b9c989-pkmss\" (UID: \"2ecf9e90-37fe-416d-9e7a-400373cfbc8d\") " pod="openshift-operators/perses-operator-5446b9c989-pkmss" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.660824 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q55v9\" (UniqueName: \"kubernetes.io/projected/2ecf9e90-37fe-416d-9e7a-400373cfbc8d-kube-api-access-q55v9\") pod \"perses-operator-5446b9c989-pkmss\" (UID: \"2ecf9e90-37fe-416d-9e7a-400373cfbc8d\") " pod="openshift-operators/perses-operator-5446b9c989-pkmss" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.661990 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/2ecf9e90-37fe-416d-9e7a-400373cfbc8d-openshift-service-ca\") pod \"perses-operator-5446b9c989-pkmss\" (UID: \"2ecf9e90-37fe-416d-9e7a-400373cfbc8d\") " pod="openshift-operators/perses-operator-5446b9c989-pkmss" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.681558 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q55v9\" (UniqueName: \"kubernetes.io/projected/2ecf9e90-37fe-416d-9e7a-400373cfbc8d-kube-api-access-q55v9\") pod \"perses-operator-5446b9c989-pkmss\" (UID: \"2ecf9e90-37fe-416d-9e7a-400373cfbc8d\") " pod="openshift-operators/perses-operator-5446b9c989-pkmss" Dec 09 09:57:19 crc kubenswrapper[4824]: I1209 09:57:19.721369 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-pkmss" Dec 09 09:57:19 crc kubenswrapper[4824]: E1209 09:57:19.752186 4824 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-pkmss_openshift-operators_2ecf9e90-37fe-416d-9e7a-400373cfbc8d_0(d1988a5fc0c6532fcb171380e6ea085af91b252d6d3976575515566f006684ee): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 09:57:19 crc kubenswrapper[4824]: E1209 09:57:19.752308 4824 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-pkmss_openshift-operators_2ecf9e90-37fe-416d-9e7a-400373cfbc8d_0(d1988a5fc0c6532fcb171380e6ea085af91b252d6d3976575515566f006684ee): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-pkmss" Dec 09 09:57:19 crc kubenswrapper[4824]: E1209 09:57:19.752372 4824 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-pkmss_openshift-operators_2ecf9e90-37fe-416d-9e7a-400373cfbc8d_0(d1988a5fc0c6532fcb171380e6ea085af91b252d6d3976575515566f006684ee): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-pkmss" Dec 09 09:57:19 crc kubenswrapper[4824]: E1209 09:57:19.752470 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-pkmss_openshift-operators(2ecf9e90-37fe-416d-9e7a-400373cfbc8d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-pkmss_openshift-operators(2ecf9e90-37fe-416d-9e7a-400373cfbc8d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-pkmss_openshift-operators_2ecf9e90-37fe-416d-9e7a-400373cfbc8d_0(d1988a5fc0c6532fcb171380e6ea085af91b252d6d3976575515566f006684ee): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-pkmss" podUID="2ecf9e90-37fe-416d-9e7a-400373cfbc8d" Dec 09 09:57:20 crc kubenswrapper[4824]: I1209 09:57:20.270685 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" event={"ID":"34793d58-1049-48ff-b8e7-187bdf42e550","Type":"ContainerStarted","Data":"a3019da7155999c81a80cad34fb8cf3bf18b564cdb27617a1ae3535a98a95a98"} Dec 09 09:57:20 crc kubenswrapper[4824]: I1209 09:57:20.271053 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:20 crc kubenswrapper[4824]: I1209 09:57:20.271164 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:20 crc kubenswrapper[4824]: I1209 09:57:20.271177 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:20 crc kubenswrapper[4824]: I1209 09:57:20.309926 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:20 crc kubenswrapper[4824]: I1209 09:57:20.310190 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:20 crc kubenswrapper[4824]: I1209 09:57:20.314940 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" podStartSLOduration=10.314914265 podStartE2EDuration="10.314914265s" podCreationTimestamp="2025-12-09 09:57:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:57:20.310797235 +0000 UTC m=+596.645301912" watchObservedRunningTime="2025-12-09 09:57:20.314914265 +0000 UTC m=+596.649418932" Dec 09 09:57:20 crc kubenswrapper[4824]: I1209 09:57:20.910745 4824 scope.go:117] "RemoveContainer" containerID="bf4f9aab454720c3762d743e8301467299a78c44fb18b35f40c349f4d60497d7" Dec 09 09:57:20 crc kubenswrapper[4824]: E1209 09:57:20.911059 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-pvksg_openshift-multus(74aa3cd7-0abc-4fc9-9697-e23b693a7611)\"" pod="openshift-multus/multus-pvksg" podUID="74aa3cd7-0abc-4fc9-9697-e23b693a7611" Dec 09 09:57:22 crc kubenswrapper[4824]: I1209 09:57:22.610327 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-qn2qz"] Dec 09 09:57:22 crc kubenswrapper[4824]: I1209 09:57:22.612026 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 09:57:22 crc kubenswrapper[4824]: I1209 09:57:22.612567 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 09:57:22 crc kubenswrapper[4824]: I1209 09:57:22.756666 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp"] Dec 09 09:57:22 crc kubenswrapper[4824]: I1209 09:57:22.756999 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" Dec 09 09:57:22 crc kubenswrapper[4824]: I1209 09:57:22.757680 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" Dec 09 09:57:22 crc kubenswrapper[4824]: E1209 09:57:22.758011 4824 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-qn2qz_openshift-operators_612ebf47-a6c4-4a62-8d97-91a003f49c44_0(539a1bfb765e44b76b5fb2a2dd9564eec145e53830489cbf1be9f0040528642f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 09:57:22 crc kubenswrapper[4824]: E1209 09:57:22.758133 4824 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-qn2qz_openshift-operators_612ebf47-a6c4-4a62-8d97-91a003f49c44_0(539a1bfb765e44b76b5fb2a2dd9564eec145e53830489cbf1be9f0040528642f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 09:57:22 crc kubenswrapper[4824]: E1209 09:57:22.758170 4824 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-qn2qz_openshift-operators_612ebf47-a6c4-4a62-8d97-91a003f49c44_0(539a1bfb765e44b76b5fb2a2dd9564eec145e53830489cbf1be9f0040528642f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 09:57:22 crc kubenswrapper[4824]: E1209 09:57:22.758241 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-qn2qz_openshift-operators(612ebf47-a6c4-4a62-8d97-91a003f49c44)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-qn2qz_openshift-operators(612ebf47-a6c4-4a62-8d97-91a003f49c44)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-qn2qz_openshift-operators_612ebf47-a6c4-4a62-8d97-91a003f49c44_0(539a1bfb765e44b76b5fb2a2dd9564eec145e53830489cbf1be9f0040528642f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" podUID="612ebf47-a6c4-4a62-8d97-91a003f49c44" Dec 09 09:57:22 crc kubenswrapper[4824]: I1209 09:57:22.766237 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf"] Dec 09 09:57:22 crc kubenswrapper[4824]: I1209 09:57:22.766384 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf" Dec 09 09:57:22 crc kubenswrapper[4824]: I1209 09:57:22.766888 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf" Dec 09 09:57:22 crc kubenswrapper[4824]: I1209 09:57:22.798099 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-xb2z8"] Dec 09 09:57:22 crc kubenswrapper[4824]: I1209 09:57:22.798241 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-xb2z8" Dec 09 09:57:22 crc kubenswrapper[4824]: I1209 09:57:22.798749 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-xb2z8" Dec 09 09:57:22 crc kubenswrapper[4824]: I1209 09:57:22.942254 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-pkmss"] Dec 09 09:57:22 crc kubenswrapper[4824]: I1209 09:57:22.942731 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-pkmss" Dec 09 09:57:22 crc kubenswrapper[4824]: I1209 09:57:22.943553 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-pkmss" Dec 09 09:57:23 crc kubenswrapper[4824]: E1209 09:57:23.391132 4824 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf_openshift-operators_eb813060-c1d8-433f-be2e-e85d102499d9_0(47f7b35d228b0ccdccea3b9e88c1bd924087088cd0b000d0fc3eb1fcfaa07d8b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 09:57:23 crc kubenswrapper[4824]: E1209 09:57:23.391666 4824 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf_openshift-operators_eb813060-c1d8-433f-be2e-e85d102499d9_0(47f7b35d228b0ccdccea3b9e88c1bd924087088cd0b000d0fc3eb1fcfaa07d8b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf" Dec 09 09:57:23 crc kubenswrapper[4824]: E1209 09:57:23.391707 4824 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf_openshift-operators_eb813060-c1d8-433f-be2e-e85d102499d9_0(47f7b35d228b0ccdccea3b9e88c1bd924087088cd0b000d0fc3eb1fcfaa07d8b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf" Dec 09 09:57:23 crc kubenswrapper[4824]: E1209 09:57:23.391811 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf_openshift-operators(eb813060-c1d8-433f-be2e-e85d102499d9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf_openshift-operators(eb813060-c1d8-433f-be2e-e85d102499d9)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf_openshift-operators_eb813060-c1d8-433f-be2e-e85d102499d9_0(47f7b35d228b0ccdccea3b9e88c1bd924087088cd0b000d0fc3eb1fcfaa07d8b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf" podUID="eb813060-c1d8-433f-be2e-e85d102499d9" Dec 09 09:57:23 crc kubenswrapper[4824]: E1209 09:57:23.398800 4824 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp_openshift-operators_cdadfe90-b4b1-4c57-a187-f6e2cbb72103_0(3a4cad2ef957de216fe78e66f7eb77b5e54c943e090909471726606bc2c05ced): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 09:57:23 crc kubenswrapper[4824]: E1209 09:57:23.398933 4824 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp_openshift-operators_cdadfe90-b4b1-4c57-a187-f6e2cbb72103_0(3a4cad2ef957de216fe78e66f7eb77b5e54c943e090909471726606bc2c05ced): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" Dec 09 09:57:23 crc kubenswrapper[4824]: E1209 09:57:23.398970 4824 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp_openshift-operators_cdadfe90-b4b1-4c57-a187-f6e2cbb72103_0(3a4cad2ef957de216fe78e66f7eb77b5e54c943e090909471726606bc2c05ced): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" Dec 09 09:57:23 crc kubenswrapper[4824]: E1209 09:57:23.399042 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp_openshift-operators(cdadfe90-b4b1-4c57-a187-f6e2cbb72103)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp_openshift-operators(cdadfe90-b4b1-4c57-a187-f6e2cbb72103)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp_openshift-operators_cdadfe90-b4b1-4c57-a187-f6e2cbb72103_0(3a4cad2ef957de216fe78e66f7eb77b5e54c943e090909471726606bc2c05ced): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" podUID="cdadfe90-b4b1-4c57-a187-f6e2cbb72103" Dec 09 09:57:23 crc kubenswrapper[4824]: E1209 09:57:23.410825 4824 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-xb2z8_openshift-operators_ca1c7fd5-ccf0-453d-9426-d77e591d729e_0(513ac05003a76fca965c71cc7389bce9cb634b9a70ba56aa5c9b0e964ca797ec): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 09:57:23 crc kubenswrapper[4824]: E1209 09:57:23.410938 4824 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-xb2z8_openshift-operators_ca1c7fd5-ccf0-453d-9426-d77e591d729e_0(513ac05003a76fca965c71cc7389bce9cb634b9a70ba56aa5c9b0e964ca797ec): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-xb2z8" Dec 09 09:57:23 crc kubenswrapper[4824]: E1209 09:57:23.410961 4824 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-xb2z8_openshift-operators_ca1c7fd5-ccf0-453d-9426-d77e591d729e_0(513ac05003a76fca965c71cc7389bce9cb634b9a70ba56aa5c9b0e964ca797ec): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-xb2z8" Dec 09 09:57:23 crc kubenswrapper[4824]: E1209 09:57:23.411003 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-xb2z8_openshift-operators(ca1c7fd5-ccf0-453d-9426-d77e591d729e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-xb2z8_openshift-operators(ca1c7fd5-ccf0-453d-9426-d77e591d729e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-xb2z8_openshift-operators_ca1c7fd5-ccf0-453d-9426-d77e591d729e_0(513ac05003a76fca965c71cc7389bce9cb634b9a70ba56aa5c9b0e964ca797ec): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-xb2z8" podUID="ca1c7fd5-ccf0-453d-9426-d77e591d729e" Dec 09 09:57:23 crc kubenswrapper[4824]: E1209 09:57:23.465367 4824 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-pkmss_openshift-operators_2ecf9e90-37fe-416d-9e7a-400373cfbc8d_0(a5d6633bdfed02e49ec241e603c771c133fde4c814c15f8c89be1b6193257b31): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 09:57:23 crc kubenswrapper[4824]: E1209 09:57:23.465484 4824 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-pkmss_openshift-operators_2ecf9e90-37fe-416d-9e7a-400373cfbc8d_0(a5d6633bdfed02e49ec241e603c771c133fde4c814c15f8c89be1b6193257b31): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-pkmss" Dec 09 09:57:23 crc kubenswrapper[4824]: E1209 09:57:23.465535 4824 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-pkmss_openshift-operators_2ecf9e90-37fe-416d-9e7a-400373cfbc8d_0(a5d6633bdfed02e49ec241e603c771c133fde4c814c15f8c89be1b6193257b31): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-pkmss" Dec 09 09:57:23 crc kubenswrapper[4824]: E1209 09:57:23.465605 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-pkmss_openshift-operators(2ecf9e90-37fe-416d-9e7a-400373cfbc8d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-pkmss_openshift-operators(2ecf9e90-37fe-416d-9e7a-400373cfbc8d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-pkmss_openshift-operators_2ecf9e90-37fe-416d-9e7a-400373cfbc8d_0(a5d6633bdfed02e49ec241e603c771c133fde4c814c15f8c89be1b6193257b31): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-pkmss" podUID="2ecf9e90-37fe-416d-9e7a-400373cfbc8d" Dec 09 09:57:31 crc kubenswrapper[4824]: I1209 09:57:31.910489 4824 scope.go:117] "RemoveContainer" containerID="bf4f9aab454720c3762d743e8301467299a78c44fb18b35f40c349f4d60497d7" Dec 09 09:57:32 crc kubenswrapper[4824]: I1209 09:57:32.611462 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-pvksg_74aa3cd7-0abc-4fc9-9697-e23b693a7611/kube-multus/2.log" Dec 09 09:57:32 crc kubenswrapper[4824]: I1209 09:57:32.611960 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pvksg" event={"ID":"74aa3cd7-0abc-4fc9-9697-e23b693a7611","Type":"ContainerStarted","Data":"5df4aa649226120ec66a437f0c55652c226a29cbc23b7f29e47c71654afe88d4"} Dec 09 09:57:32 crc kubenswrapper[4824]: I1209 09:57:32.866095 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 09:57:32 crc kubenswrapper[4824]: I1209 09:57:32.866184 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 09:57:32 crc kubenswrapper[4824]: I1209 09:57:32.910374 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 09:57:32 crc kubenswrapper[4824]: I1209 09:57:32.911401 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 09:57:32 crc kubenswrapper[4824]: E1209 09:57:32.976156 4824 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-qn2qz_openshift-operators_612ebf47-a6c4-4a62-8d97-91a003f49c44_0(e51328d26278c82c9628b39fab052fddfddb86b606301e2f5021d91280e53144): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 09:57:32 crc kubenswrapper[4824]: E1209 09:57:32.976649 4824 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-qn2qz_openshift-operators_612ebf47-a6c4-4a62-8d97-91a003f49c44_0(e51328d26278c82c9628b39fab052fddfddb86b606301e2f5021d91280e53144): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 09:57:32 crc kubenswrapper[4824]: E1209 09:57:32.976676 4824 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-qn2qz_openshift-operators_612ebf47-a6c4-4a62-8d97-91a003f49c44_0(e51328d26278c82c9628b39fab052fddfddb86b606301e2f5021d91280e53144): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 09:57:32 crc kubenswrapper[4824]: E1209 09:57:32.976742 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-qn2qz_openshift-operators(612ebf47-a6c4-4a62-8d97-91a003f49c44)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-qn2qz_openshift-operators(612ebf47-a6c4-4a62-8d97-91a003f49c44)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-qn2qz_openshift-operators_612ebf47-a6c4-4a62-8d97-91a003f49c44_0(e51328d26278c82c9628b39fab052fddfddb86b606301e2f5021d91280e53144): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" podUID="612ebf47-a6c4-4a62-8d97-91a003f49c44" Dec 09 09:57:33 crc kubenswrapper[4824]: I1209 09:57:33.917243 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" Dec 09 09:57:33 crc kubenswrapper[4824]: I1209 09:57:33.918805 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" Dec 09 09:57:34 crc kubenswrapper[4824]: E1209 09:57:34.019115 4824 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp_openshift-operators_cdadfe90-b4b1-4c57-a187-f6e2cbb72103_0(90657afcc8200da33513546cf54d7480d77989a59581ce618055ba0d30290634): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 09:57:34 crc kubenswrapper[4824]: E1209 09:57:34.019242 4824 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp_openshift-operators_cdadfe90-b4b1-4c57-a187-f6e2cbb72103_0(90657afcc8200da33513546cf54d7480d77989a59581ce618055ba0d30290634): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" Dec 09 09:57:34 crc kubenswrapper[4824]: E1209 09:57:34.019277 4824 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp_openshift-operators_cdadfe90-b4b1-4c57-a187-f6e2cbb72103_0(90657afcc8200da33513546cf54d7480d77989a59581ce618055ba0d30290634): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" Dec 09 09:57:34 crc kubenswrapper[4824]: E1209 09:57:34.019353 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp_openshift-operators(cdadfe90-b4b1-4c57-a187-f6e2cbb72103)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp_openshift-operators(cdadfe90-b4b1-4c57-a187-f6e2cbb72103)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp_openshift-operators_cdadfe90-b4b1-4c57-a187-f6e2cbb72103_0(90657afcc8200da33513546cf54d7480d77989a59581ce618055ba0d30290634): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" podUID="cdadfe90-b4b1-4c57-a187-f6e2cbb72103" Dec 09 09:57:35 crc kubenswrapper[4824]: I1209 09:57:35.910041 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-pkmss" Dec 09 09:57:35 crc kubenswrapper[4824]: I1209 09:57:35.910855 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-pkmss" Dec 09 09:57:36 crc kubenswrapper[4824]: I1209 09:57:36.415473 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-pkmss"] Dec 09 09:57:36 crc kubenswrapper[4824]: I1209 09:57:36.641605 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-pkmss" event={"ID":"2ecf9e90-37fe-416d-9e7a-400373cfbc8d","Type":"ContainerStarted","Data":"1bdd459b6c94cd25c37c359593018df4943296fc4045dbc5a49e41491a60b331"} Dec 09 09:57:37 crc kubenswrapper[4824]: I1209 09:57:37.914872 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-xb2z8" Dec 09 09:57:37 crc kubenswrapper[4824]: I1209 09:57:37.915135 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-xb2z8" Dec 09 09:57:38 crc kubenswrapper[4824]: I1209 09:57:38.522813 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-xb2z8"] Dec 09 09:57:38 crc kubenswrapper[4824]: I1209 09:57:38.672548 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-xb2z8" event={"ID":"ca1c7fd5-ccf0-453d-9426-d77e591d729e","Type":"ContainerStarted","Data":"2c978fe1606142f2fe80685f51ba084b92322cdc0e322537500c55d1cbbef883"} Dec 09 09:57:38 crc kubenswrapper[4824]: I1209 09:57:38.909906 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf" Dec 09 09:57:38 crc kubenswrapper[4824]: I1209 09:57:38.911209 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf" Dec 09 09:57:39 crc kubenswrapper[4824]: I1209 09:57:39.438012 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf"] Dec 09 09:57:39 crc kubenswrapper[4824]: I1209 09:57:39.693572 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf" event={"ID":"eb813060-c1d8-433f-be2e-e85d102499d9","Type":"ContainerStarted","Data":"a514c4300e310be51977635d5b15ab9c856eb3ae522b5621275e22e40e34d9d1"} Dec 09 09:57:41 crc kubenswrapper[4824]: I1209 09:57:41.295203 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" Dec 09 09:57:44 crc kubenswrapper[4824]: I1209 09:57:44.909341 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" Dec 09 09:57:44 crc kubenswrapper[4824]: I1209 09:57:44.910285 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" Dec 09 09:57:46 crc kubenswrapper[4824]: I1209 09:57:46.910335 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 09:57:46 crc kubenswrapper[4824]: I1209 09:57:46.910582 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 09:57:51 crc kubenswrapper[4824]: E1209 09:57:51.883040 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:9aec4c328ec43e40481e06ca5808deead74b75c0aacb90e9e72966c3fa14f385" Dec 09 09:57:51 crc kubenswrapper[4824]: E1209 09:57:51.884078 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:perses-operator,Image:registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:9aec4c328ec43e40481e06ca5808deead74b75c0aacb90e9e72966c3fa14f385,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{134217728 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openshift-service-ca,ReadOnly:true,MountPath:/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-q55v9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod perses-operator-5446b9c989-pkmss_openshift-operators(2ecf9e90-37fe-416d-9e7a-400373cfbc8d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 09:57:51 crc kubenswrapper[4824]: E1209 09:57:51.885438 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/perses-operator-5446b9c989-pkmss" podUID="2ecf9e90-37fe-416d-9e7a-400373cfbc8d" Dec 09 09:57:52 crc kubenswrapper[4824]: E1209 09:57:52.901415 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:9aec4c328ec43e40481e06ca5808deead74b75c0aacb90e9e72966c3fa14f385\\\"\"" pod="openshift-operators/perses-operator-5446b9c989-pkmss" podUID="2ecf9e90-37fe-416d-9e7a-400373cfbc8d" Dec 09 09:57:53 crc kubenswrapper[4824]: I1209 09:57:53.324852 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-qn2qz"] Dec 09 09:57:53 crc kubenswrapper[4824]: W1209 09:57:53.331739 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod612ebf47_a6c4_4a62_8d97_91a003f49c44.slice/crio-9fe1f27e8c7e7cea81ba362d6ed165c9ec8ce5c4ea6537ea33991c96ac1a9fff WatchSource:0}: Error finding container 9fe1f27e8c7e7cea81ba362d6ed165c9ec8ce5c4ea6537ea33991c96ac1a9fff: Status 404 returned error can't find the container with id 9fe1f27e8c7e7cea81ba362d6ed165c9ec8ce5c4ea6537ea33991c96ac1a9fff Dec 09 09:57:53 crc kubenswrapper[4824]: I1209 09:57:53.466656 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp"] Dec 09 09:57:53 crc kubenswrapper[4824]: W1209 09:57:53.472811 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcdadfe90_b4b1_4c57_a187_f6e2cbb72103.slice/crio-b71a1cbf62ef134f2ec894d1e304f012d9ad00ee16db67f5aea21cdb9d449abf WatchSource:0}: Error finding container b71a1cbf62ef134f2ec894d1e304f012d9ad00ee16db67f5aea21cdb9d449abf: Status 404 returned error can't find the container with id b71a1cbf62ef134f2ec894d1e304f012d9ad00ee16db67f5aea21cdb9d449abf Dec 09 09:57:53 crc kubenswrapper[4824]: I1209 09:57:53.928340 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" event={"ID":"cdadfe90-b4b1-4c57-a187-f6e2cbb72103","Type":"ContainerStarted","Data":"30b68a2348d55354ba0442efccb13d9d6f4f49e5b0c2b3762a95b3e37402894a"} Dec 09 09:57:53 crc kubenswrapper[4824]: I1209 09:57:53.928391 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" event={"ID":"cdadfe90-b4b1-4c57-a187-f6e2cbb72103","Type":"ContainerStarted","Data":"b71a1cbf62ef134f2ec894d1e304f012d9ad00ee16db67f5aea21cdb9d449abf"} Dec 09 09:57:53 crc kubenswrapper[4824]: I1209 09:57:53.928404 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-xb2z8" event={"ID":"ca1c7fd5-ccf0-453d-9426-d77e591d729e","Type":"ContainerStarted","Data":"7e39a8c661cf8208f7ada3427a4c0534b7ebd03b9aca0327be87565de5a2ea44"} Dec 09 09:57:53 crc kubenswrapper[4824]: I1209 09:57:53.928609 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" event={"ID":"612ebf47-a6c4-4a62-8d97-91a003f49c44","Type":"ContainerStarted","Data":"9fe1f27e8c7e7cea81ba362d6ed165c9ec8ce5c4ea6537ea33991c96ac1a9fff"} Dec 09 09:57:53 crc kubenswrapper[4824]: I1209 09:57:53.937677 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf" event={"ID":"eb813060-c1d8-433f-be2e-e85d102499d9","Type":"ContainerStarted","Data":"4119bd3e75982040b5be696a19fdd073e93235fd29ff3da8066ad89d91d52814"} Dec 09 09:57:53 crc kubenswrapper[4824]: I1209 09:57:53.980794 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf" podStartSLOduration=22.405489006 podStartE2EDuration="35.980753705s" podCreationTimestamp="2025-12-09 09:57:18 +0000 UTC" firstStartedPulling="2025-12-09 09:57:39.450356515 +0000 UTC m=+615.784861192" lastFinishedPulling="2025-12-09 09:57:53.025621234 +0000 UTC m=+629.360125891" observedRunningTime="2025-12-09 09:57:53.978511685 +0000 UTC m=+630.313016362" watchObservedRunningTime="2025-12-09 09:57:53.980753705 +0000 UTC m=+630.315258372" Dec 09 09:57:54 crc kubenswrapper[4824]: I1209 09:57:54.023294 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-xb2z8" podStartSLOduration=21.544993726 podStartE2EDuration="36.023265795s" podCreationTimestamp="2025-12-09 09:57:18 +0000 UTC" firstStartedPulling="2025-12-09 09:57:38.566237181 +0000 UTC m=+614.900741848" lastFinishedPulling="2025-12-09 09:57:53.04450925 +0000 UTC m=+629.379013917" observedRunningTime="2025-12-09 09:57:54.005766624 +0000 UTC m=+630.340271301" watchObservedRunningTime="2025-12-09 09:57:54.023265795 +0000 UTC m=+630.357770462" Dec 09 09:57:54 crc kubenswrapper[4824]: I1209 09:57:54.033355 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp" podStartSLOduration=36.033330722 podStartE2EDuration="36.033330722s" podCreationTimestamp="2025-12-09 09:57:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 09:57:54.030916046 +0000 UTC m=+630.365420713" watchObservedRunningTime="2025-12-09 09:57:54.033330722 +0000 UTC m=+630.367835389" Dec 09 09:57:59 crc kubenswrapper[4824]: I1209 09:57:59.986652 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" event={"ID":"612ebf47-a6c4-4a62-8d97-91a003f49c44","Type":"ContainerStarted","Data":"6211e2b6e456aecff8e6484685cb1dbfb8ed11c2a63a1ec44ca688b1a82a5c22"} Dec 09 09:57:59 crc kubenswrapper[4824]: I1209 09:57:59.987353 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 09:57:59 crc kubenswrapper[4824]: I1209 09:57:59.992393 4824 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-qn2qz container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.5:8081/healthz\": dial tcp 10.217.0.5:8081: connect: connection refused" start-of-body= Dec 09 09:57:59 crc kubenswrapper[4824]: I1209 09:57:59.992460 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" podUID="612ebf47-a6c4-4a62-8d97-91a003f49c44" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.5:8081/healthz\": dial tcp 10.217.0.5:8081: connect: connection refused" Dec 09 09:58:00 crc kubenswrapper[4824]: I1209 09:58:00.018680 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" podStartSLOduration=34.674129152 podStartE2EDuration="41.018658518s" podCreationTimestamp="2025-12-09 09:57:19 +0000 UTC" firstStartedPulling="2025-12-09 09:57:53.335419668 +0000 UTC m=+629.669924335" lastFinishedPulling="2025-12-09 09:57:59.679949034 +0000 UTC m=+636.014453701" observedRunningTime="2025-12-09 09:58:00.006556447 +0000 UTC m=+636.341061134" watchObservedRunningTime="2025-12-09 09:58:00.018658518 +0000 UTC m=+636.353163185" Dec 09 09:58:01 crc kubenswrapper[4824]: I1209 09:58:01.256325 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 09:58:02 crc kubenswrapper[4824]: I1209 09:58:02.861009 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 09:58:02 crc kubenswrapper[4824]: I1209 09:58:02.861105 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 09:58:02 crc kubenswrapper[4824]: I1209 09:58:02.861184 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 09:58:02 crc kubenswrapper[4824]: I1209 09:58:02.862420 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b9242f271427aa606711a2a58ca0863fdbf70e4a738de9f900b42d848af928bc"} pod="openshift-machine-config-operator/machine-config-daemon-dth8x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 09:58:02 crc kubenswrapper[4824]: I1209 09:58:02.862559 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" containerID="cri-o://b9242f271427aa606711a2a58ca0863fdbf70e4a738de9f900b42d848af928bc" gracePeriod=600 Dec 09 09:58:03 crc kubenswrapper[4824]: I1209 09:58:03.007833 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerID="b9242f271427aa606711a2a58ca0863fdbf70e4a738de9f900b42d848af928bc" exitCode=0 Dec 09 09:58:03 crc kubenswrapper[4824]: I1209 09:58:03.007904 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerDied","Data":"b9242f271427aa606711a2a58ca0863fdbf70e4a738de9f900b42d848af928bc"} Dec 09 09:58:03 crc kubenswrapper[4824]: I1209 09:58:03.008214 4824 scope.go:117] "RemoveContainer" containerID="ea7720fbe70782cdf3ecb70ae54c936954803442a5c7d2c6bc34bd0e635bfbfe" Dec 09 09:58:04 crc kubenswrapper[4824]: I1209 09:58:04.018824 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerStarted","Data":"ef93ca94e26d32583c5e2e8c602aa45c24cbdf5e3b1e4f117c890d24188f2377"} Dec 09 09:58:07 crc kubenswrapper[4824]: I1209 09:58:07.041943 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-pkmss" event={"ID":"2ecf9e90-37fe-416d-9e7a-400373cfbc8d","Type":"ContainerStarted","Data":"a1002ad43243314fa1ba439a6f691b4b26ed9ca04e6d46c30e9f9274180e62f9"} Dec 09 09:58:07 crc kubenswrapper[4824]: I1209 09:58:07.042629 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-pkmss" Dec 09 09:58:07 crc kubenswrapper[4824]: I1209 09:58:07.067222 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-pkmss" podStartSLOduration=17.807140754 podStartE2EDuration="48.067192082s" podCreationTimestamp="2025-12-09 09:57:19 +0000 UTC" firstStartedPulling="2025-12-09 09:57:36.428344486 +0000 UTC m=+612.762849153" lastFinishedPulling="2025-12-09 09:58:06.688395814 +0000 UTC m=+643.022900481" observedRunningTime="2025-12-09 09:58:07.060149949 +0000 UTC m=+643.394654636" watchObservedRunningTime="2025-12-09 09:58:07.067192082 +0000 UTC m=+643.401696749" Dec 09 09:58:11 crc kubenswrapper[4824]: I1209 09:58:11.863815 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-d4569"] Dec 09 09:58:11 crc kubenswrapper[4824]: I1209 09:58:11.865295 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-d4569" Dec 09 09:58:11 crc kubenswrapper[4824]: I1209 09:58:11.874528 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 09 09:58:11 crc kubenswrapper[4824]: I1209 09:58:11.874718 4824 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-ccjl5" Dec 09 09:58:11 crc kubenswrapper[4824]: I1209 09:58:11.875474 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 09 09:58:11 crc kubenswrapper[4824]: I1209 09:58:11.879012 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-2zl4f"] Dec 09 09:58:11 crc kubenswrapper[4824]: I1209 09:58:11.880495 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-2zl4f" Dec 09 09:58:11 crc kubenswrapper[4824]: I1209 09:58:11.884699 4824 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-bqsf8" Dec 09 09:58:11 crc kubenswrapper[4824]: I1209 09:58:11.887427 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-2zl4f"] Dec 09 09:58:11 crc kubenswrapper[4824]: I1209 09:58:11.893542 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-d4569"] Dec 09 09:58:11 crc kubenswrapper[4824]: I1209 09:58:11.904550 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-dbnmx"] Dec 09 09:58:11 crc kubenswrapper[4824]: I1209 09:58:11.905089 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjx7j\" (UniqueName: \"kubernetes.io/projected/c35e598c-4ffa-4c4b-bdec-9940fe209d32-kube-api-access-pjx7j\") pod \"cert-manager-cainjector-7f985d654d-d4569\" (UID: \"c35e598c-4ffa-4c4b-bdec-9940fe209d32\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-d4569" Dec 09 09:58:11 crc kubenswrapper[4824]: I1209 09:58:11.905160 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kc42\" (UniqueName: \"kubernetes.io/projected/a3e087c0-4d4b-4678-8a71-1cf7ee33a867-kube-api-access-7kc42\") pod \"cert-manager-5b446d88c5-2zl4f\" (UID: \"a3e087c0-4d4b-4678-8a71-1cf7ee33a867\") " pod="cert-manager/cert-manager-5b446d88c5-2zl4f" Dec 09 09:58:11 crc kubenswrapper[4824]: I1209 09:58:11.905751 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" Dec 09 09:58:11 crc kubenswrapper[4824]: I1209 09:58:11.913214 4824 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-flzr9" Dec 09 09:58:11 crc kubenswrapper[4824]: I1209 09:58:11.940900 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-dbnmx"] Dec 09 09:58:12 crc kubenswrapper[4824]: I1209 09:58:12.006068 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjx7j\" (UniqueName: \"kubernetes.io/projected/c35e598c-4ffa-4c4b-bdec-9940fe209d32-kube-api-access-pjx7j\") pod \"cert-manager-cainjector-7f985d654d-d4569\" (UID: \"c35e598c-4ffa-4c4b-bdec-9940fe209d32\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-d4569" Dec 09 09:58:12 crc kubenswrapper[4824]: I1209 09:58:12.006134 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kc42\" (UniqueName: \"kubernetes.io/projected/a3e087c0-4d4b-4678-8a71-1cf7ee33a867-kube-api-access-7kc42\") pod \"cert-manager-5b446d88c5-2zl4f\" (UID: \"a3e087c0-4d4b-4678-8a71-1cf7ee33a867\") " pod="cert-manager/cert-manager-5b446d88c5-2zl4f" Dec 09 09:58:12 crc kubenswrapper[4824]: I1209 09:58:12.006161 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg5tv\" (UniqueName: \"kubernetes.io/projected/9f87669d-33a9-4269-808c-ca6c718f762c-kube-api-access-mg5tv\") pod \"cert-manager-webhook-5655c58dd6-dbnmx\" (UID: \"9f87669d-33a9-4269-808c-ca6c718f762c\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" Dec 09 09:58:12 crc kubenswrapper[4824]: I1209 09:58:12.029250 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjx7j\" (UniqueName: \"kubernetes.io/projected/c35e598c-4ffa-4c4b-bdec-9940fe209d32-kube-api-access-pjx7j\") pod \"cert-manager-cainjector-7f985d654d-d4569\" (UID: \"c35e598c-4ffa-4c4b-bdec-9940fe209d32\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-d4569" Dec 09 09:58:12 crc kubenswrapper[4824]: I1209 09:58:12.029484 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kc42\" (UniqueName: \"kubernetes.io/projected/a3e087c0-4d4b-4678-8a71-1cf7ee33a867-kube-api-access-7kc42\") pod \"cert-manager-5b446d88c5-2zl4f\" (UID: \"a3e087c0-4d4b-4678-8a71-1cf7ee33a867\") " pod="cert-manager/cert-manager-5b446d88c5-2zl4f" Dec 09 09:58:12 crc kubenswrapper[4824]: I1209 09:58:12.107928 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mg5tv\" (UniqueName: \"kubernetes.io/projected/9f87669d-33a9-4269-808c-ca6c718f762c-kube-api-access-mg5tv\") pod \"cert-manager-webhook-5655c58dd6-dbnmx\" (UID: \"9f87669d-33a9-4269-808c-ca6c718f762c\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" Dec 09 09:58:12 crc kubenswrapper[4824]: I1209 09:58:12.127514 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mg5tv\" (UniqueName: \"kubernetes.io/projected/9f87669d-33a9-4269-808c-ca6c718f762c-kube-api-access-mg5tv\") pod \"cert-manager-webhook-5655c58dd6-dbnmx\" (UID: \"9f87669d-33a9-4269-808c-ca6c718f762c\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" Dec 09 09:58:12 crc kubenswrapper[4824]: I1209 09:58:12.187808 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-d4569" Dec 09 09:58:12 crc kubenswrapper[4824]: I1209 09:58:12.203607 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-2zl4f" Dec 09 09:58:12 crc kubenswrapper[4824]: I1209 09:58:12.232676 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" Dec 09 09:58:12 crc kubenswrapper[4824]: I1209 09:58:12.821715 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-d4569"] Dec 09 09:58:12 crc kubenswrapper[4824]: W1209 09:58:12.833704 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc35e598c_4ffa_4c4b_bdec_9940fe209d32.slice/crio-947a9468c5030721fc09b2e03dfd9c2e1c9bb353a272a604de281811b714e606 WatchSource:0}: Error finding container 947a9468c5030721fc09b2e03dfd9c2e1c9bb353a272a604de281811b714e606: Status 404 returned error can't find the container with id 947a9468c5030721fc09b2e03dfd9c2e1c9bb353a272a604de281811b714e606 Dec 09 09:58:12 crc kubenswrapper[4824]: I1209 09:58:12.875965 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-2zl4f"] Dec 09 09:58:13 crc kubenswrapper[4824]: I1209 09:58:13.099871 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-d4569" event={"ID":"c35e598c-4ffa-4c4b-bdec-9940fe209d32","Type":"ContainerStarted","Data":"947a9468c5030721fc09b2e03dfd9c2e1c9bb353a272a604de281811b714e606"} Dec 09 09:58:13 crc kubenswrapper[4824]: I1209 09:58:13.101201 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-2zl4f" event={"ID":"a3e087c0-4d4b-4678-8a71-1cf7ee33a867","Type":"ContainerStarted","Data":"682fb1113432f1f64bbb1a574cd7b961bed5108e0a40d0d61fe29c65c86e6187"} Dec 09 09:58:13 crc kubenswrapper[4824]: I1209 09:58:13.127699 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-dbnmx"] Dec 09 09:58:13 crc kubenswrapper[4824]: W1209 09:58:13.132622 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f87669d_33a9_4269_808c_ca6c718f762c.slice/crio-2e3f4bc81e6f58caede0c65b44e5abf2921cd09095238036eece0a1c3cfe912d WatchSource:0}: Error finding container 2e3f4bc81e6f58caede0c65b44e5abf2921cd09095238036eece0a1c3cfe912d: Status 404 returned error can't find the container with id 2e3f4bc81e6f58caede0c65b44e5abf2921cd09095238036eece0a1c3cfe912d Dec 09 09:58:14 crc kubenswrapper[4824]: I1209 09:58:14.113315 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" event={"ID":"9f87669d-33a9-4269-808c-ca6c718f762c","Type":"ContainerStarted","Data":"2e3f4bc81e6f58caede0c65b44e5abf2921cd09095238036eece0a1c3cfe912d"} Dec 09 09:58:17 crc kubenswrapper[4824]: I1209 09:58:17.137690 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-d4569" event={"ID":"c35e598c-4ffa-4c4b-bdec-9940fe209d32","Type":"ContainerStarted","Data":"2fa2923a15394de9c23674e396cb9410998e6b4ecb75ab81b2c338d8deb21e6b"} Dec 09 09:58:17 crc kubenswrapper[4824]: I1209 09:58:17.140085 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" event={"ID":"9f87669d-33a9-4269-808c-ca6c718f762c","Type":"ContainerStarted","Data":"1d23dc9a9237a313ad67f6e0183bda8cbae96cfe627b77a3827352d7702bd4b6"} Dec 09 09:58:17 crc kubenswrapper[4824]: I1209 09:58:17.140260 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" Dec 09 09:58:17 crc kubenswrapper[4824]: I1209 09:58:17.166596 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-d4569" podStartSLOduration=2.435802627 podStartE2EDuration="6.16655888s" podCreationTimestamp="2025-12-09 09:58:11 +0000 UTC" firstStartedPulling="2025-12-09 09:58:12.836707566 +0000 UTC m=+649.171212233" lastFinishedPulling="2025-12-09 09:58:16.567463819 +0000 UTC m=+652.901968486" observedRunningTime="2025-12-09 09:58:17.157935128 +0000 UTC m=+653.492439815" watchObservedRunningTime="2025-12-09 09:58:17.16655888 +0000 UTC m=+653.501063547" Dec 09 09:58:17 crc kubenswrapper[4824]: I1209 09:58:17.175227 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" podStartSLOduration=2.705765935 podStartE2EDuration="6.175204692s" podCreationTimestamp="2025-12-09 09:58:11 +0000 UTC" firstStartedPulling="2025-12-09 09:58:13.13698452 +0000 UTC m=+649.471489197" lastFinishedPulling="2025-12-09 09:58:16.606423287 +0000 UTC m=+652.940927954" observedRunningTime="2025-12-09 09:58:17.175111229 +0000 UTC m=+653.509615906" watchObservedRunningTime="2025-12-09 09:58:17.175204692 +0000 UTC m=+653.509709359" Dec 09 09:58:19 crc kubenswrapper[4824]: I1209 09:58:19.725856 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-pkmss" Dec 09 09:58:21 crc kubenswrapper[4824]: I1209 09:58:21.183131 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-2zl4f" event={"ID":"a3e087c0-4d4b-4678-8a71-1cf7ee33a867","Type":"ContainerStarted","Data":"740e1d6aee2ab462081c7942189a8f48e27a801dcf915c74633fc2981cc155ae"} Dec 09 09:58:21 crc kubenswrapper[4824]: I1209 09:58:21.203923 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-2zl4f" podStartSLOduration=3.09784851 podStartE2EDuration="10.203888885s" podCreationTimestamp="2025-12-09 09:58:11 +0000 UTC" firstStartedPulling="2025-12-09 09:58:12.897089289 +0000 UTC m=+649.231593956" lastFinishedPulling="2025-12-09 09:58:20.003129664 +0000 UTC m=+656.337634331" observedRunningTime="2025-12-09 09:58:21.200124787 +0000 UTC m=+657.534629464" watchObservedRunningTime="2025-12-09 09:58:21.203888885 +0000 UTC m=+657.538393552" Dec 09 09:58:22 crc kubenswrapper[4824]: I1209 09:58:22.235603 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" Dec 09 09:58:48 crc kubenswrapper[4824]: I1209 09:58:48.163335 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz"] Dec 09 09:58:48 crc kubenswrapper[4824]: I1209 09:58:48.166359 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz" Dec 09 09:58:48 crc kubenswrapper[4824]: I1209 09:58:48.172461 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 09 09:58:48 crc kubenswrapper[4824]: I1209 09:58:48.190043 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz"] Dec 09 09:58:48 crc kubenswrapper[4824]: I1209 09:58:48.266697 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bf54b053-f1e7-4b01-90f8-2c0aac92c4b4-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz\" (UID: \"bf54b053-f1e7-4b01-90f8-2c0aac92c4b4\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz" Dec 09 09:58:48 crc kubenswrapper[4824]: I1209 09:58:48.266837 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bf54b053-f1e7-4b01-90f8-2c0aac92c4b4-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz\" (UID: \"bf54b053-f1e7-4b01-90f8-2c0aac92c4b4\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz" Dec 09 09:58:48 crc kubenswrapper[4824]: I1209 09:58:48.266928 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j62s\" (UniqueName: \"kubernetes.io/projected/bf54b053-f1e7-4b01-90f8-2c0aac92c4b4-kube-api-access-2j62s\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz\" (UID: \"bf54b053-f1e7-4b01-90f8-2c0aac92c4b4\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz" Dec 09 09:58:48 crc kubenswrapper[4824]: I1209 09:58:48.368720 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bf54b053-f1e7-4b01-90f8-2c0aac92c4b4-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz\" (UID: \"bf54b053-f1e7-4b01-90f8-2c0aac92c4b4\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz" Dec 09 09:58:48 crc kubenswrapper[4824]: I1209 09:58:48.368897 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j62s\" (UniqueName: \"kubernetes.io/projected/bf54b053-f1e7-4b01-90f8-2c0aac92c4b4-kube-api-access-2j62s\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz\" (UID: \"bf54b053-f1e7-4b01-90f8-2c0aac92c4b4\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz" Dec 09 09:58:48 crc kubenswrapper[4824]: I1209 09:58:48.369031 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bf54b053-f1e7-4b01-90f8-2c0aac92c4b4-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz\" (UID: \"bf54b053-f1e7-4b01-90f8-2c0aac92c4b4\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz" Dec 09 09:58:48 crc kubenswrapper[4824]: I1209 09:58:48.369418 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bf54b053-f1e7-4b01-90f8-2c0aac92c4b4-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz\" (UID: \"bf54b053-f1e7-4b01-90f8-2c0aac92c4b4\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz" Dec 09 09:58:48 crc kubenswrapper[4824]: I1209 09:58:48.369634 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bf54b053-f1e7-4b01-90f8-2c0aac92c4b4-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz\" (UID: \"bf54b053-f1e7-4b01-90f8-2c0aac92c4b4\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz" Dec 09 09:58:48 crc kubenswrapper[4824]: I1209 09:58:48.396326 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j62s\" (UniqueName: \"kubernetes.io/projected/bf54b053-f1e7-4b01-90f8-2c0aac92c4b4-kube-api-access-2j62s\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz\" (UID: \"bf54b053-f1e7-4b01-90f8-2c0aac92c4b4\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz" Dec 09 09:58:48 crc kubenswrapper[4824]: I1209 09:58:48.494664 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz" Dec 09 09:58:49 crc kubenswrapper[4824]: I1209 09:58:48.999982 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz"] Dec 09 09:58:49 crc kubenswrapper[4824]: I1209 09:58:49.405054 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz" event={"ID":"bf54b053-f1e7-4b01-90f8-2c0aac92c4b4","Type":"ContainerStarted","Data":"e71017df202b3b0d5ad1fe25f1982948b748fb1782d10f349f285c5817b6625a"} Dec 09 09:58:49 crc kubenswrapper[4824]: I1209 09:58:49.405537 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz" event={"ID":"bf54b053-f1e7-4b01-90f8-2c0aac92c4b4","Type":"ContainerStarted","Data":"138d066c7a21267a153f13351f88138a804d9531bda3fa8a2200b7c563dc388e"} Dec 09 09:58:50 crc kubenswrapper[4824]: I1209 09:58:50.413699 4824 generic.go:334] "Generic (PLEG): container finished" podID="bf54b053-f1e7-4b01-90f8-2c0aac92c4b4" containerID="e71017df202b3b0d5ad1fe25f1982948b748fb1782d10f349f285c5817b6625a" exitCode=0 Dec 09 09:58:50 crc kubenswrapper[4824]: I1209 09:58:50.413774 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz" event={"ID":"bf54b053-f1e7-4b01-90f8-2c0aac92c4b4","Type":"ContainerDied","Data":"e71017df202b3b0d5ad1fe25f1982948b748fb1782d10f349f285c5817b6625a"} Dec 09 09:58:53 crc kubenswrapper[4824]: I1209 09:58:53.440927 4824 generic.go:334] "Generic (PLEG): container finished" podID="bf54b053-f1e7-4b01-90f8-2c0aac92c4b4" containerID="f532dc791c6d7441b3372dcf87aa527ff91495373f56dade64dbdc6076223a90" exitCode=0 Dec 09 09:58:53 crc kubenswrapper[4824]: I1209 09:58:53.441097 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz" event={"ID":"bf54b053-f1e7-4b01-90f8-2c0aac92c4b4","Type":"ContainerDied","Data":"f532dc791c6d7441b3372dcf87aa527ff91495373f56dade64dbdc6076223a90"} Dec 09 09:58:54 crc kubenswrapper[4824]: I1209 09:58:54.461498 4824 generic.go:334] "Generic (PLEG): container finished" podID="bf54b053-f1e7-4b01-90f8-2c0aac92c4b4" containerID="74dfa742733c4bb2bff96345d590f7df4e6353754f5e3415d735ed6ca9ed31ea" exitCode=0 Dec 09 09:58:54 crc kubenswrapper[4824]: I1209 09:58:54.461572 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz" event={"ID":"bf54b053-f1e7-4b01-90f8-2c0aac92c4b4","Type":"ContainerDied","Data":"74dfa742733c4bb2bff96345d590f7df4e6353754f5e3415d735ed6ca9ed31ea"} Dec 09 09:58:54 crc kubenswrapper[4824]: I1209 09:58:54.565705 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5"] Dec 09 09:58:54 crc kubenswrapper[4824]: I1209 09:58:54.568332 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5" Dec 09 09:58:54 crc kubenswrapper[4824]: I1209 09:58:54.577676 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5"] Dec 09 09:58:54 crc kubenswrapper[4824]: I1209 09:58:54.675700 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cf722168-25bd-4faa-9d35-b59240bb718f-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5\" (UID: \"cf722168-25bd-4faa-9d35-b59240bb718f\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5" Dec 09 09:58:54 crc kubenswrapper[4824]: I1209 09:58:54.675867 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cf722168-25bd-4faa-9d35-b59240bb718f-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5\" (UID: \"cf722168-25bd-4faa-9d35-b59240bb718f\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5" Dec 09 09:58:54 crc kubenswrapper[4824]: I1209 09:58:54.675965 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpb6j\" (UniqueName: \"kubernetes.io/projected/cf722168-25bd-4faa-9d35-b59240bb718f-kube-api-access-kpb6j\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5\" (UID: \"cf722168-25bd-4faa-9d35-b59240bb718f\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5" Dec 09 09:58:54 crc kubenswrapper[4824]: I1209 09:58:54.778966 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpb6j\" (UniqueName: \"kubernetes.io/projected/cf722168-25bd-4faa-9d35-b59240bb718f-kube-api-access-kpb6j\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5\" (UID: \"cf722168-25bd-4faa-9d35-b59240bb718f\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5" Dec 09 09:58:54 crc kubenswrapper[4824]: I1209 09:58:54.779077 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cf722168-25bd-4faa-9d35-b59240bb718f-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5\" (UID: \"cf722168-25bd-4faa-9d35-b59240bb718f\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5" Dec 09 09:58:54 crc kubenswrapper[4824]: I1209 09:58:54.779130 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cf722168-25bd-4faa-9d35-b59240bb718f-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5\" (UID: \"cf722168-25bd-4faa-9d35-b59240bb718f\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5" Dec 09 09:58:54 crc kubenswrapper[4824]: I1209 09:58:54.779651 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cf722168-25bd-4faa-9d35-b59240bb718f-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5\" (UID: \"cf722168-25bd-4faa-9d35-b59240bb718f\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5" Dec 09 09:58:54 crc kubenswrapper[4824]: I1209 09:58:54.779738 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cf722168-25bd-4faa-9d35-b59240bb718f-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5\" (UID: \"cf722168-25bd-4faa-9d35-b59240bb718f\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5" Dec 09 09:58:54 crc kubenswrapper[4824]: I1209 09:58:54.801564 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpb6j\" (UniqueName: \"kubernetes.io/projected/cf722168-25bd-4faa-9d35-b59240bb718f-kube-api-access-kpb6j\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5\" (UID: \"cf722168-25bd-4faa-9d35-b59240bb718f\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5" Dec 09 09:58:54 crc kubenswrapper[4824]: I1209 09:58:54.891092 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5" Dec 09 09:58:55 crc kubenswrapper[4824]: I1209 09:58:55.118776 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5"] Dec 09 09:58:55 crc kubenswrapper[4824]: W1209 09:58:55.127167 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf722168_25bd_4faa_9d35_b59240bb718f.slice/crio-ccb9141fe182efc6a520dcee34efd91063d6eba1019a46c0f393118cab7270a2 WatchSource:0}: Error finding container ccb9141fe182efc6a520dcee34efd91063d6eba1019a46c0f393118cab7270a2: Status 404 returned error can't find the container with id ccb9141fe182efc6a520dcee34efd91063d6eba1019a46c0f393118cab7270a2 Dec 09 09:58:55 crc kubenswrapper[4824]: I1209 09:58:55.472454 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5" event={"ID":"cf722168-25bd-4faa-9d35-b59240bb718f","Type":"ContainerStarted","Data":"40461bbd0ae8a8ce54c6be6f3bd6283638f8701d44c17151dfcd656c22c534e7"} Dec 09 09:58:55 crc kubenswrapper[4824]: I1209 09:58:55.472595 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5" event={"ID":"cf722168-25bd-4faa-9d35-b59240bb718f","Type":"ContainerStarted","Data":"ccb9141fe182efc6a520dcee34efd91063d6eba1019a46c0f393118cab7270a2"} Dec 09 09:58:56 crc kubenswrapper[4824]: I1209 09:58:56.668145 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz" Dec 09 09:58:56 crc kubenswrapper[4824]: I1209 09:58:56.838413 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bf54b053-f1e7-4b01-90f8-2c0aac92c4b4-bundle\") pod \"bf54b053-f1e7-4b01-90f8-2c0aac92c4b4\" (UID: \"bf54b053-f1e7-4b01-90f8-2c0aac92c4b4\") " Dec 09 09:58:56 crc kubenswrapper[4824]: I1209 09:58:56.838526 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2j62s\" (UniqueName: \"kubernetes.io/projected/bf54b053-f1e7-4b01-90f8-2c0aac92c4b4-kube-api-access-2j62s\") pod \"bf54b053-f1e7-4b01-90f8-2c0aac92c4b4\" (UID: \"bf54b053-f1e7-4b01-90f8-2c0aac92c4b4\") " Dec 09 09:58:56 crc kubenswrapper[4824]: I1209 09:58:56.838747 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bf54b053-f1e7-4b01-90f8-2c0aac92c4b4-util\") pod \"bf54b053-f1e7-4b01-90f8-2c0aac92c4b4\" (UID: \"bf54b053-f1e7-4b01-90f8-2c0aac92c4b4\") " Dec 09 09:58:56 crc kubenswrapper[4824]: I1209 09:58:56.840016 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf54b053-f1e7-4b01-90f8-2c0aac92c4b4-bundle" (OuterVolumeSpecName: "bundle") pod "bf54b053-f1e7-4b01-90f8-2c0aac92c4b4" (UID: "bf54b053-f1e7-4b01-90f8-2c0aac92c4b4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:58:56 crc kubenswrapper[4824]: I1209 09:58:56.840701 4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bf54b053-f1e7-4b01-90f8-2c0aac92c4b4-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 09:58:56 crc kubenswrapper[4824]: I1209 09:58:56.846109 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf54b053-f1e7-4b01-90f8-2c0aac92c4b4-kube-api-access-2j62s" (OuterVolumeSpecName: "kube-api-access-2j62s") pod "bf54b053-f1e7-4b01-90f8-2c0aac92c4b4" (UID: "bf54b053-f1e7-4b01-90f8-2c0aac92c4b4"). InnerVolumeSpecName "kube-api-access-2j62s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:58:56 crc kubenswrapper[4824]: I1209 09:58:56.854203 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf54b053-f1e7-4b01-90f8-2c0aac92c4b4-util" (OuterVolumeSpecName: "util") pod "bf54b053-f1e7-4b01-90f8-2c0aac92c4b4" (UID: "bf54b053-f1e7-4b01-90f8-2c0aac92c4b4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:58:56 crc kubenswrapper[4824]: I1209 09:58:56.943073 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2j62s\" (UniqueName: \"kubernetes.io/projected/bf54b053-f1e7-4b01-90f8-2c0aac92c4b4-kube-api-access-2j62s\") on node \"crc\" DevicePath \"\"" Dec 09 09:58:56 crc kubenswrapper[4824]: I1209 09:58:56.943137 4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bf54b053-f1e7-4b01-90f8-2c0aac92c4b4-util\") on node \"crc\" DevicePath \"\"" Dec 09 09:58:57 crc kubenswrapper[4824]: I1209 09:58:57.490174 4824 generic.go:334] "Generic (PLEG): container finished" podID="cf722168-25bd-4faa-9d35-b59240bb718f" containerID="40461bbd0ae8a8ce54c6be6f3bd6283638f8701d44c17151dfcd656c22c534e7" exitCode=0 Dec 09 09:58:57 crc kubenswrapper[4824]: I1209 09:58:57.490266 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5" event={"ID":"cf722168-25bd-4faa-9d35-b59240bb718f","Type":"ContainerDied","Data":"40461bbd0ae8a8ce54c6be6f3bd6283638f8701d44c17151dfcd656c22c534e7"} Dec 09 09:58:57 crc kubenswrapper[4824]: I1209 09:58:57.494198 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz" event={"ID":"bf54b053-f1e7-4b01-90f8-2c0aac92c4b4","Type":"ContainerDied","Data":"138d066c7a21267a153f13351f88138a804d9531bda3fa8a2200b7c563dc388e"} Dec 09 09:58:57 crc kubenswrapper[4824]: I1209 09:58:57.494252 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="138d066c7a21267a153f13351f88138a804d9531bda3fa8a2200b7c563dc388e" Dec 09 09:58:57 crc kubenswrapper[4824]: I1209 09:58:57.494276 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz" Dec 09 09:58:59 crc kubenswrapper[4824]: I1209 09:58:59.512908 4824 generic.go:334] "Generic (PLEG): container finished" podID="cf722168-25bd-4faa-9d35-b59240bb718f" containerID="f5e37f466c19cf45be0d8ace7bf908033e6ad37f5b2f9e37607d14818c5ef807" exitCode=0 Dec 09 09:58:59 crc kubenswrapper[4824]: I1209 09:58:59.512991 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5" event={"ID":"cf722168-25bd-4faa-9d35-b59240bb718f","Type":"ContainerDied","Data":"f5e37f466c19cf45be0d8ace7bf908033e6ad37f5b2f9e37607d14818c5ef807"} Dec 09 09:59:00 crc kubenswrapper[4824]: I1209 09:59:00.524898 4824 generic.go:334] "Generic (PLEG): container finished" podID="cf722168-25bd-4faa-9d35-b59240bb718f" containerID="e9b04dda10b1c6eb3ca3dadabfcb2b8909f0938c68842653ff2e8f5d157081e5" exitCode=0 Dec 09 09:59:00 crc kubenswrapper[4824]: I1209 09:59:00.525004 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5" event={"ID":"cf722168-25bd-4faa-9d35-b59240bb718f","Type":"ContainerDied","Data":"e9b04dda10b1c6eb3ca3dadabfcb2b8909f0938c68842653ff2e8f5d157081e5"} Dec 09 09:59:02 crc kubenswrapper[4824]: I1209 09:59:02.048886 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5" Dec 09 09:59:02 crc kubenswrapper[4824]: I1209 09:59:02.250572 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cf722168-25bd-4faa-9d35-b59240bb718f-util\") pod \"cf722168-25bd-4faa-9d35-b59240bb718f\" (UID: \"cf722168-25bd-4faa-9d35-b59240bb718f\") " Dec 09 09:59:02 crc kubenswrapper[4824]: I1209 09:59:02.250654 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cf722168-25bd-4faa-9d35-b59240bb718f-bundle\") pod \"cf722168-25bd-4faa-9d35-b59240bb718f\" (UID: \"cf722168-25bd-4faa-9d35-b59240bb718f\") " Dec 09 09:59:02 crc kubenswrapper[4824]: I1209 09:59:02.250750 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpb6j\" (UniqueName: \"kubernetes.io/projected/cf722168-25bd-4faa-9d35-b59240bb718f-kube-api-access-kpb6j\") pod \"cf722168-25bd-4faa-9d35-b59240bb718f\" (UID: \"cf722168-25bd-4faa-9d35-b59240bb718f\") " Dec 09 09:59:02 crc kubenswrapper[4824]: I1209 09:59:02.253200 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf722168-25bd-4faa-9d35-b59240bb718f-bundle" (OuterVolumeSpecName: "bundle") pod "cf722168-25bd-4faa-9d35-b59240bb718f" (UID: "cf722168-25bd-4faa-9d35-b59240bb718f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:59:02 crc kubenswrapper[4824]: I1209 09:59:02.258440 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf722168-25bd-4faa-9d35-b59240bb718f-kube-api-access-kpb6j" (OuterVolumeSpecName: "kube-api-access-kpb6j") pod "cf722168-25bd-4faa-9d35-b59240bb718f" (UID: "cf722168-25bd-4faa-9d35-b59240bb718f"). InnerVolumeSpecName "kube-api-access-kpb6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 09:59:02 crc kubenswrapper[4824]: I1209 09:59:02.263090 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf722168-25bd-4faa-9d35-b59240bb718f-util" (OuterVolumeSpecName: "util") pod "cf722168-25bd-4faa-9d35-b59240bb718f" (UID: "cf722168-25bd-4faa-9d35-b59240bb718f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 09:59:02 crc kubenswrapper[4824]: I1209 09:59:02.352924 4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cf722168-25bd-4faa-9d35-b59240bb718f-util\") on node \"crc\" DevicePath \"\"" Dec 09 09:59:02 crc kubenswrapper[4824]: I1209 09:59:02.352976 4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cf722168-25bd-4faa-9d35-b59240bb718f-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 09:59:02 crc kubenswrapper[4824]: I1209 09:59:02.352990 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpb6j\" (UniqueName: \"kubernetes.io/projected/cf722168-25bd-4faa-9d35-b59240bb718f-kube-api-access-kpb6j\") on node \"crc\" DevicePath \"\"" Dec 09 09:59:02 crc kubenswrapper[4824]: I1209 09:59:02.544963 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5" event={"ID":"cf722168-25bd-4faa-9d35-b59240bb718f","Type":"ContainerDied","Data":"ccb9141fe182efc6a520dcee34efd91063d6eba1019a46c0f393118cab7270a2"} Dec 09 09:59:02 crc kubenswrapper[4824]: I1209 09:59:02.545045 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5" Dec 09 09:59:02 crc kubenswrapper[4824]: I1209 09:59:02.546032 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccb9141fe182efc6a520dcee34efd91063d6eba1019a46c0f393118cab7270a2" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.475663 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7"] Dec 09 09:59:03 crc kubenswrapper[4824]: E1209 09:59:03.476129 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf54b053-f1e7-4b01-90f8-2c0aac92c4b4" containerName="extract" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.476150 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf54b053-f1e7-4b01-90f8-2c0aac92c4b4" containerName="extract" Dec 09 09:59:03 crc kubenswrapper[4824]: E1209 09:59:03.476187 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf722168-25bd-4faa-9d35-b59240bb718f" containerName="extract" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.476199 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf722168-25bd-4faa-9d35-b59240bb718f" containerName="extract" Dec 09 09:59:03 crc kubenswrapper[4824]: E1209 09:59:03.476210 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf722168-25bd-4faa-9d35-b59240bb718f" containerName="util" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.476221 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf722168-25bd-4faa-9d35-b59240bb718f" containerName="util" Dec 09 09:59:03 crc kubenswrapper[4824]: E1209 09:59:03.476251 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf54b053-f1e7-4b01-90f8-2c0aac92c4b4" containerName="pull" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.476262 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf54b053-f1e7-4b01-90f8-2c0aac92c4b4" containerName="pull" Dec 09 09:59:03 crc kubenswrapper[4824]: E1209 09:59:03.476276 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf722168-25bd-4faa-9d35-b59240bb718f" containerName="pull" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.476284 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf722168-25bd-4faa-9d35-b59240bb718f" containerName="pull" Dec 09 09:59:03 crc kubenswrapper[4824]: E1209 09:59:03.476305 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf54b053-f1e7-4b01-90f8-2c0aac92c4b4" containerName="util" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.476316 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf54b053-f1e7-4b01-90f8-2c0aac92c4b4" containerName="util" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.476481 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf722168-25bd-4faa-9d35-b59240bb718f" containerName="extract" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.476506 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf54b053-f1e7-4b01-90f8-2c0aac92c4b4" containerName="extract" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.479404 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.483711 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.484256 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.484444 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-sz542" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.484677 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.484987 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.485161 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.512424 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7"] Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.578703 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s549x\" (UniqueName: \"kubernetes.io/projected/352c7865-5a0b-4ebb-93f2-513ce433ff8f-kube-api-access-s549x\") pod \"loki-operator-controller-manager-6bf69c47b7-8fjw7\" (UID: \"352c7865-5a0b-4ebb-93f2-513ce433ff8f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.578979 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/352c7865-5a0b-4ebb-93f2-513ce433ff8f-manager-config\") pod \"loki-operator-controller-manager-6bf69c47b7-8fjw7\" (UID: \"352c7865-5a0b-4ebb-93f2-513ce433ff8f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.579041 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/352c7865-5a0b-4ebb-93f2-513ce433ff8f-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-6bf69c47b7-8fjw7\" (UID: \"352c7865-5a0b-4ebb-93f2-513ce433ff8f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.579076 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/352c7865-5a0b-4ebb-93f2-513ce433ff8f-webhook-cert\") pod \"loki-operator-controller-manager-6bf69c47b7-8fjw7\" (UID: \"352c7865-5a0b-4ebb-93f2-513ce433ff8f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.579103 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/352c7865-5a0b-4ebb-93f2-513ce433ff8f-apiservice-cert\") pod \"loki-operator-controller-manager-6bf69c47b7-8fjw7\" (UID: \"352c7865-5a0b-4ebb-93f2-513ce433ff8f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.680820 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/352c7865-5a0b-4ebb-93f2-513ce433ff8f-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-6bf69c47b7-8fjw7\" (UID: \"352c7865-5a0b-4ebb-93f2-513ce433ff8f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.681321 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/352c7865-5a0b-4ebb-93f2-513ce433ff8f-apiservice-cert\") pod \"loki-operator-controller-manager-6bf69c47b7-8fjw7\" (UID: \"352c7865-5a0b-4ebb-93f2-513ce433ff8f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.681346 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/352c7865-5a0b-4ebb-93f2-513ce433ff8f-webhook-cert\") pod \"loki-operator-controller-manager-6bf69c47b7-8fjw7\" (UID: \"352c7865-5a0b-4ebb-93f2-513ce433ff8f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.681501 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s549x\" (UniqueName: \"kubernetes.io/projected/352c7865-5a0b-4ebb-93f2-513ce433ff8f-kube-api-access-s549x\") pod \"loki-operator-controller-manager-6bf69c47b7-8fjw7\" (UID: \"352c7865-5a0b-4ebb-93f2-513ce433ff8f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.682070 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/352c7865-5a0b-4ebb-93f2-513ce433ff8f-manager-config\") pod \"loki-operator-controller-manager-6bf69c47b7-8fjw7\" (UID: \"352c7865-5a0b-4ebb-93f2-513ce433ff8f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.683208 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/352c7865-5a0b-4ebb-93f2-513ce433ff8f-manager-config\") pod \"loki-operator-controller-manager-6bf69c47b7-8fjw7\" (UID: \"352c7865-5a0b-4ebb-93f2-513ce433ff8f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.695715 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/352c7865-5a0b-4ebb-93f2-513ce433ff8f-apiservice-cert\") pod \"loki-operator-controller-manager-6bf69c47b7-8fjw7\" (UID: \"352c7865-5a0b-4ebb-93f2-513ce433ff8f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.696232 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/352c7865-5a0b-4ebb-93f2-513ce433ff8f-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-6bf69c47b7-8fjw7\" (UID: \"352c7865-5a0b-4ebb-93f2-513ce433ff8f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.707702 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/352c7865-5a0b-4ebb-93f2-513ce433ff8f-webhook-cert\") pod \"loki-operator-controller-manager-6bf69c47b7-8fjw7\" (UID: \"352c7865-5a0b-4ebb-93f2-513ce433ff8f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.712402 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s549x\" (UniqueName: \"kubernetes.io/projected/352c7865-5a0b-4ebb-93f2-513ce433ff8f-kube-api-access-s549x\") pod \"loki-operator-controller-manager-6bf69c47b7-8fjw7\" (UID: \"352c7865-5a0b-4ebb-93f2-513ce433ff8f\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" Dec 09 09:59:03 crc kubenswrapper[4824]: I1209 09:59:03.804917 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" Dec 09 09:59:04 crc kubenswrapper[4824]: I1209 09:59:04.312587 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7"] Dec 09 09:59:04 crc kubenswrapper[4824]: I1209 09:59:04.570321 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" event={"ID":"352c7865-5a0b-4ebb-93f2-513ce433ff8f","Type":"ContainerStarted","Data":"9fafa759d0eda2040099b57ee2d78106815197057833c54a9867fd957d764966"} Dec 09 09:59:10 crc kubenswrapper[4824]: I1209 09:59:10.550706 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-txx2m"] Dec 09 09:59:10 crc kubenswrapper[4824]: I1209 09:59:10.552247 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-txx2m" Dec 09 09:59:10 crc kubenswrapper[4824]: I1209 09:59:10.555583 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"cluster-logging-operator-dockercfg-rlbff" Dec 09 09:59:10 crc kubenswrapper[4824]: I1209 09:59:10.564423 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"kube-root-ca.crt" Dec 09 09:59:10 crc kubenswrapper[4824]: I1209 09:59:10.564493 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"openshift-service-ca.crt" Dec 09 09:59:10 crc kubenswrapper[4824]: I1209 09:59:10.722148 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rr4x\" (UniqueName: \"kubernetes.io/projected/333c3894-c96b-4511-8c1b-8ea76bae7ef6-kube-api-access-9rr4x\") pod \"cluster-logging-operator-ff9846bd-txx2m\" (UID: \"333c3894-c96b-4511-8c1b-8ea76bae7ef6\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-txx2m" Dec 09 09:59:10 crc kubenswrapper[4824]: I1209 09:59:10.751023 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-txx2m"] Dec 09 09:59:10 crc kubenswrapper[4824]: I1209 09:59:10.824635 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rr4x\" (UniqueName: \"kubernetes.io/projected/333c3894-c96b-4511-8c1b-8ea76bae7ef6-kube-api-access-9rr4x\") pod \"cluster-logging-operator-ff9846bd-txx2m\" (UID: \"333c3894-c96b-4511-8c1b-8ea76bae7ef6\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-txx2m" Dec 09 09:59:10 crc kubenswrapper[4824]: I1209 09:59:10.855387 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rr4x\" (UniqueName: \"kubernetes.io/projected/333c3894-c96b-4511-8c1b-8ea76bae7ef6-kube-api-access-9rr4x\") pod \"cluster-logging-operator-ff9846bd-txx2m\" (UID: \"333c3894-c96b-4511-8c1b-8ea76bae7ef6\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-txx2m" Dec 09 09:59:10 crc kubenswrapper[4824]: I1209 09:59:10.881353 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-txx2m" Dec 09 09:59:14 crc kubenswrapper[4824]: I1209 09:59:14.292530 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-txx2m"] Dec 09 09:59:14 crc kubenswrapper[4824]: W1209 09:59:14.300372 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod333c3894_c96b_4511_8c1b_8ea76bae7ef6.slice/crio-ea3dd8cfb2a26d580cbc4616dac6371bd66857bc3ef1a4cd5a4cc7f4a62a52ea WatchSource:0}: Error finding container ea3dd8cfb2a26d580cbc4616dac6371bd66857bc3ef1a4cd5a4cc7f4a62a52ea: Status 404 returned error can't find the container with id ea3dd8cfb2a26d580cbc4616dac6371bd66857bc3ef1a4cd5a4cc7f4a62a52ea Dec 09 09:59:14 crc kubenswrapper[4824]: I1209 09:59:14.789977 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" event={"ID":"352c7865-5a0b-4ebb-93f2-513ce433ff8f","Type":"ContainerStarted","Data":"e427734a99512a3fca2d9ff58d913920a92fe006a0a467df372b3bea14f49369"} Dec 09 09:59:14 crc kubenswrapper[4824]: I1209 09:59:14.792249 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-txx2m" event={"ID":"333c3894-c96b-4511-8c1b-8ea76bae7ef6","Type":"ContainerStarted","Data":"ea3dd8cfb2a26d580cbc4616dac6371bd66857bc3ef1a4cd5a4cc7f4a62a52ea"} Dec 09 09:59:30 crc kubenswrapper[4824]: E1209 09:59:30.805001 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f15ebe396f96093861d528a3307a3e38ac2d4dff594f793c1e56011a3a909175" Dec 09 09:59:30 crc kubenswrapper[4824]: E1209 09:59:30.806225 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cluster-logging-operator,Image:registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f15ebe396f96093861d528a3307a3e38ac2d4dff594f793c1e56011a3a909175,Command:[cluster-logging-operator],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.annotations['olm.targetNamespaces'],},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:cluster-logging-operator,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_VECTOR,Value:registry.redhat.io/openshift-logging/vector-rhel9@sha256:438aa27c0408214bc64d01b20f233d698fa48344aa2dd878bc090232f227c17c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_LOG_FILE_METRIC_EXPORTER,Value:registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:d4512c0a403fa3cddc646e566f5b6c77ac17c0558a08d8f99ffcdafaba9fba3a,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-logging.v6.2.6,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9rr4x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000690000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cluster-logging-operator-ff9846bd-txx2m_openshift-logging(333c3894-c96b-4511-8c1b-8ea76bae7ef6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 09:59:30 crc kubenswrapper[4824]: E1209 09:59:30.807511 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cluster-logging-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-logging/cluster-logging-operator-ff9846bd-txx2m" podUID="333c3894-c96b-4511-8c1b-8ea76bae7ef6" Dec 09 09:59:31 crc kubenswrapper[4824]: E1209 09:59:31.437105 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cluster-logging-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f15ebe396f96093861d528a3307a3e38ac2d4dff594f793c1e56011a3a909175\\\"\"" pod="openshift-logging/cluster-logging-operator-ff9846bd-txx2m" podUID="333c3894-c96b-4511-8c1b-8ea76bae7ef6" Dec 09 09:59:32 crc kubenswrapper[4824]: E1209 09:59:32.091715 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24" Dec 09 09:59:32 crc kubenswrapper[4824]: E1209 09:59:32.091979 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --tls-cert-file=/var/run/secrets/serving-cert/tls.crt --tls-private-key-file=/var/run/secrets/serving-cert/tls.key --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256,TLS_RSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_256_GCM_SHA384,TLS_RSA_WITH_AES_128_CBC_SHA256 --tls-min-version=VersionTLS12 --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:loki-operator.v6.2.6,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:loki-operator-metrics-cert,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s549x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000700000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod loki-operator-controller-manager-6bf69c47b7-8fjw7_openshift-operators-redhat(352c7865-5a0b-4ebb-93f2-513ce433ff8f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 09:59:32 crc kubenswrapper[4824]: E1209 09:59:32.093204 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" podUID="352c7865-5a0b-4ebb-93f2-513ce433ff8f" Dec 09 09:59:32 crc kubenswrapper[4824]: I1209 09:59:32.440623 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" Dec 09 09:59:32 crc kubenswrapper[4824]: E1209 09:59:32.442974 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24\\\"\"" pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" podUID="352c7865-5a0b-4ebb-93f2-513ce433ff8f" Dec 09 09:59:32 crc kubenswrapper[4824]: I1209 09:59:32.447375 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" Dec 09 09:59:33 crc kubenswrapper[4824]: E1209 09:59:33.449637 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24\\\"\"" pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" podUID="352c7865-5a0b-4ebb-93f2-513ce433ff8f" Dec 09 09:59:34 crc kubenswrapper[4824]: E1209 09:59:34.455261 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24\\\"\"" pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" podUID="352c7865-5a0b-4ebb-93f2-513ce433ff8f" Dec 09 09:59:46 crc kubenswrapper[4824]: I1209 09:59:46.544362 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-txx2m" event={"ID":"333c3894-c96b-4511-8c1b-8ea76bae7ef6","Type":"ContainerStarted","Data":"fc890c30612051d22093ef9884146370f9bf0f819a9d948434b9acae825e74ba"} Dec 09 09:59:46 crc kubenswrapper[4824]: I1209 09:59:46.568696 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/cluster-logging-operator-ff9846bd-txx2m" podStartSLOduration=4.687865314 podStartE2EDuration="36.56866876s" podCreationTimestamp="2025-12-09 09:59:10 +0000 UTC" firstStartedPulling="2025-12-09 09:59:14.306724282 +0000 UTC m=+710.641228949" lastFinishedPulling="2025-12-09 09:59:46.187527728 +0000 UTC m=+742.522032395" observedRunningTime="2025-12-09 09:59:46.567219434 +0000 UTC m=+742.901724111" watchObservedRunningTime="2025-12-09 09:59:46.56866876 +0000 UTC m=+742.903173437" Dec 09 09:59:49 crc kubenswrapper[4824]: I1209 09:59:49.600479 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" event={"ID":"352c7865-5a0b-4ebb-93f2-513ce433ff8f","Type":"ContainerStarted","Data":"14bd9def8a01c7ee609b2e20c0cd5f94e258b066263ccab45319d31e259ad029"} Dec 09 09:59:49 crc kubenswrapper[4824]: I1209 09:59:49.631833 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" podStartSLOduration=2.508594311 podStartE2EDuration="46.631804344s" podCreationTimestamp="2025-12-09 09:59:03 +0000 UTC" firstStartedPulling="2025-12-09 09:59:04.326309871 +0000 UTC m=+700.660814538" lastFinishedPulling="2025-12-09 09:59:48.449519904 +0000 UTC m=+744.784024571" observedRunningTime="2025-12-09 09:59:49.625079801 +0000 UTC m=+745.959584478" watchObservedRunningTime="2025-12-09 09:59:49.631804344 +0000 UTC m=+745.966309021" Dec 09 09:59:55 crc kubenswrapper[4824]: I1209 09:59:55.117024 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Dec 09 09:59:55 crc kubenswrapper[4824]: I1209 09:59:55.126450 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Dec 09 09:59:55 crc kubenswrapper[4824]: I1209 09:59:55.129513 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Dec 09 09:59:55 crc kubenswrapper[4824]: I1209 09:59:55.129885 4824 reflector.go:368] Caches populated for *v1.Secret from object-"minio-dev"/"default-dockercfg-6rz8x" Dec 09 09:59:55 crc kubenswrapper[4824]: I1209 09:59:55.130587 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Dec 09 09:59:55 crc kubenswrapper[4824]: I1209 09:59:55.135975 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Dec 09 09:59:55 crc kubenswrapper[4824]: I1209 09:59:55.306168 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr8vw\" (UniqueName: \"kubernetes.io/projected/d9cee2cc-bb90-469d-8c72-29cf32fd4305-kube-api-access-tr8vw\") pod \"minio\" (UID: \"d9cee2cc-bb90-469d-8c72-29cf32fd4305\") " pod="minio-dev/minio" Dec 09 09:59:55 crc kubenswrapper[4824]: I1209 09:59:55.306751 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2a9cdda6-1ddc-4f43-858e-57b10b411bdc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2a9cdda6-1ddc-4f43-858e-57b10b411bdc\") pod \"minio\" (UID: \"d9cee2cc-bb90-469d-8c72-29cf32fd4305\") " pod="minio-dev/minio" Dec 09 09:59:55 crc kubenswrapper[4824]: I1209 09:59:55.407999 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr8vw\" (UniqueName: \"kubernetes.io/projected/d9cee2cc-bb90-469d-8c72-29cf32fd4305-kube-api-access-tr8vw\") pod \"minio\" (UID: \"d9cee2cc-bb90-469d-8c72-29cf32fd4305\") " pod="minio-dev/minio" Dec 09 09:59:55 crc kubenswrapper[4824]: I1209 09:59:55.408423 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2a9cdda6-1ddc-4f43-858e-57b10b411bdc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2a9cdda6-1ddc-4f43-858e-57b10b411bdc\") pod \"minio\" (UID: \"d9cee2cc-bb90-469d-8c72-29cf32fd4305\") " pod="minio-dev/minio" Dec 09 09:59:55 crc kubenswrapper[4824]: I1209 09:59:55.422984 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 09:59:55 crc kubenswrapper[4824]: I1209 09:59:55.423040 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2a9cdda6-1ddc-4f43-858e-57b10b411bdc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2a9cdda6-1ddc-4f43-858e-57b10b411bdc\") pod \"minio\" (UID: \"d9cee2cc-bb90-469d-8c72-29cf32fd4305\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/afb693dcf505d215d71e05a51454f42b5d6d9438eba91ae6e85a53d30720d973/globalmount\"" pod="minio-dev/minio" Dec 09 09:59:55 crc kubenswrapper[4824]: I1209 09:59:55.430162 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr8vw\" (UniqueName: \"kubernetes.io/projected/d9cee2cc-bb90-469d-8c72-29cf32fd4305-kube-api-access-tr8vw\") pod \"minio\" (UID: \"d9cee2cc-bb90-469d-8c72-29cf32fd4305\") " pod="minio-dev/minio" Dec 09 09:59:55 crc kubenswrapper[4824]: I1209 09:59:55.458125 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2a9cdda6-1ddc-4f43-858e-57b10b411bdc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2a9cdda6-1ddc-4f43-858e-57b10b411bdc\") pod \"minio\" (UID: \"d9cee2cc-bb90-469d-8c72-29cf32fd4305\") " pod="minio-dev/minio" Dec 09 09:59:55 crc kubenswrapper[4824]: I1209 09:59:55.748025 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Dec 09 09:59:56 crc kubenswrapper[4824]: I1209 09:59:56.221842 4824 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 09 09:59:56 crc kubenswrapper[4824]: I1209 09:59:56.247928 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Dec 09 09:59:56 crc kubenswrapper[4824]: I1209 09:59:56.696751 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"d9cee2cc-bb90-469d-8c72-29cf32fd4305","Type":"ContainerStarted","Data":"b0e75f019f4738ba6798950361bfd8e4ea7056173fc760fd4aea8420590abf8b"} Dec 09 10:00:00 crc kubenswrapper[4824]: I1209 10:00:00.176756 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421240-cqlx2"] Dec 09 10:00:00 crc kubenswrapper[4824]: I1209 10:00:00.178869 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421240-cqlx2" Dec 09 10:00:00 crc kubenswrapper[4824]: I1209 10:00:00.181374 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 10:00:00 crc kubenswrapper[4824]: I1209 10:00:00.182170 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 10:00:00 crc kubenswrapper[4824]: I1209 10:00:00.198350 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421240-cqlx2"] Dec 09 10:00:00 crc kubenswrapper[4824]: I1209 10:00:00.296913 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ca530df-dd8f-427e-bc32-6b107395b6d5-secret-volume\") pod \"collect-profiles-29421240-cqlx2\" (UID: \"4ca530df-dd8f-427e-bc32-6b107395b6d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421240-cqlx2" Dec 09 10:00:00 crc kubenswrapper[4824]: I1209 10:00:00.296970 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnxsr\" (UniqueName: \"kubernetes.io/projected/4ca530df-dd8f-427e-bc32-6b107395b6d5-kube-api-access-nnxsr\") pod \"collect-profiles-29421240-cqlx2\" (UID: \"4ca530df-dd8f-427e-bc32-6b107395b6d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421240-cqlx2" Dec 09 10:00:00 crc kubenswrapper[4824]: I1209 10:00:00.297073 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ca530df-dd8f-427e-bc32-6b107395b6d5-config-volume\") pod \"collect-profiles-29421240-cqlx2\" (UID: \"4ca530df-dd8f-427e-bc32-6b107395b6d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421240-cqlx2" Dec 09 10:00:00 crc kubenswrapper[4824]: I1209 10:00:00.399204 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ca530df-dd8f-427e-bc32-6b107395b6d5-secret-volume\") pod \"collect-profiles-29421240-cqlx2\" (UID: \"4ca530df-dd8f-427e-bc32-6b107395b6d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421240-cqlx2" Dec 09 10:00:00 crc kubenswrapper[4824]: I1209 10:00:00.399267 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnxsr\" (UniqueName: \"kubernetes.io/projected/4ca530df-dd8f-427e-bc32-6b107395b6d5-kube-api-access-nnxsr\") pod \"collect-profiles-29421240-cqlx2\" (UID: \"4ca530df-dd8f-427e-bc32-6b107395b6d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421240-cqlx2" Dec 09 10:00:00 crc kubenswrapper[4824]: I1209 10:00:00.399327 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ca530df-dd8f-427e-bc32-6b107395b6d5-config-volume\") pod \"collect-profiles-29421240-cqlx2\" (UID: \"4ca530df-dd8f-427e-bc32-6b107395b6d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421240-cqlx2" Dec 09 10:00:00 crc kubenswrapper[4824]: I1209 10:00:00.400996 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ca530df-dd8f-427e-bc32-6b107395b6d5-config-volume\") pod \"collect-profiles-29421240-cqlx2\" (UID: \"4ca530df-dd8f-427e-bc32-6b107395b6d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421240-cqlx2" Dec 09 10:00:00 crc kubenswrapper[4824]: I1209 10:00:00.419239 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ca530df-dd8f-427e-bc32-6b107395b6d5-secret-volume\") pod \"collect-profiles-29421240-cqlx2\" (UID: \"4ca530df-dd8f-427e-bc32-6b107395b6d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421240-cqlx2" Dec 09 10:00:00 crc kubenswrapper[4824]: I1209 10:00:00.424214 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnxsr\" (UniqueName: \"kubernetes.io/projected/4ca530df-dd8f-427e-bc32-6b107395b6d5-kube-api-access-nnxsr\") pod \"collect-profiles-29421240-cqlx2\" (UID: \"4ca530df-dd8f-427e-bc32-6b107395b6d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421240-cqlx2" Dec 09 10:00:00 crc kubenswrapper[4824]: I1209 10:00:00.541595 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421240-cqlx2" Dec 09 10:00:01 crc kubenswrapper[4824]: I1209 10:00:01.053616 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421240-cqlx2"] Dec 09 10:00:01 crc kubenswrapper[4824]: I1209 10:00:01.742168 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421240-cqlx2" event={"ID":"4ca530df-dd8f-427e-bc32-6b107395b6d5","Type":"ContainerStarted","Data":"448b6595c82b3a5d4f873fbd2d468b683458cbd4e6c76699fac8e8505395051b"} Dec 09 10:00:01 crc kubenswrapper[4824]: I1209 10:00:01.742250 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421240-cqlx2" event={"ID":"4ca530df-dd8f-427e-bc32-6b107395b6d5","Type":"ContainerStarted","Data":"f079a0b24a09cd84dbb836f8896a1ddd826fba8e289a48331059aeb8dee3eb0b"} Dec 09 10:00:01 crc kubenswrapper[4824]: I1209 10:00:01.774317 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29421240-cqlx2" podStartSLOduration=1.774290321 podStartE2EDuration="1.774290321s" podCreationTimestamp="2025-12-09 10:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:00:01.765635318 +0000 UTC m=+758.100140015" watchObservedRunningTime="2025-12-09 10:00:01.774290321 +0000 UTC m=+758.108794988" Dec 09 10:00:02 crc kubenswrapper[4824]: I1209 10:00:02.758415 4824 generic.go:334] "Generic (PLEG): container finished" podID="4ca530df-dd8f-427e-bc32-6b107395b6d5" containerID="448b6595c82b3a5d4f873fbd2d468b683458cbd4e6c76699fac8e8505395051b" exitCode=0 Dec 09 10:00:02 crc kubenswrapper[4824]: I1209 10:00:02.758498 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421240-cqlx2" event={"ID":"4ca530df-dd8f-427e-bc32-6b107395b6d5","Type":"ContainerDied","Data":"448b6595c82b3a5d4f873fbd2d468b683458cbd4e6c76699fac8e8505395051b"} Dec 09 10:00:06 crc kubenswrapper[4824]: I1209 10:00:06.576231 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421240-cqlx2" Dec 09 10:00:06 crc kubenswrapper[4824]: I1209 10:00:06.777192 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ca530df-dd8f-427e-bc32-6b107395b6d5-config-volume\") pod \"4ca530df-dd8f-427e-bc32-6b107395b6d5\" (UID: \"4ca530df-dd8f-427e-bc32-6b107395b6d5\") " Dec 09 10:00:06 crc kubenswrapper[4824]: I1209 10:00:06.777289 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnxsr\" (UniqueName: \"kubernetes.io/projected/4ca530df-dd8f-427e-bc32-6b107395b6d5-kube-api-access-nnxsr\") pod \"4ca530df-dd8f-427e-bc32-6b107395b6d5\" (UID: \"4ca530df-dd8f-427e-bc32-6b107395b6d5\") " Dec 09 10:00:06 crc kubenswrapper[4824]: I1209 10:00:06.777473 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ca530df-dd8f-427e-bc32-6b107395b6d5-secret-volume\") pod \"4ca530df-dd8f-427e-bc32-6b107395b6d5\" (UID: \"4ca530df-dd8f-427e-bc32-6b107395b6d5\") " Dec 09 10:00:06 crc kubenswrapper[4824]: I1209 10:00:06.779184 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ca530df-dd8f-427e-bc32-6b107395b6d5-config-volume" (OuterVolumeSpecName: "config-volume") pod "4ca530df-dd8f-427e-bc32-6b107395b6d5" (UID: "4ca530df-dd8f-427e-bc32-6b107395b6d5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:00:06 crc kubenswrapper[4824]: I1209 10:00:06.790556 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ca530df-dd8f-427e-bc32-6b107395b6d5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4ca530df-dd8f-427e-bc32-6b107395b6d5" (UID: "4ca530df-dd8f-427e-bc32-6b107395b6d5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:00:06 crc kubenswrapper[4824]: I1209 10:00:06.793008 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ca530df-dd8f-427e-bc32-6b107395b6d5-kube-api-access-nnxsr" (OuterVolumeSpecName: "kube-api-access-nnxsr") pod "4ca530df-dd8f-427e-bc32-6b107395b6d5" (UID: "4ca530df-dd8f-427e-bc32-6b107395b6d5"). InnerVolumeSpecName "kube-api-access-nnxsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:00:06 crc kubenswrapper[4824]: I1209 10:00:06.794638 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421240-cqlx2" event={"ID":"4ca530df-dd8f-427e-bc32-6b107395b6d5","Type":"ContainerDied","Data":"f079a0b24a09cd84dbb836f8896a1ddd826fba8e289a48331059aeb8dee3eb0b"} Dec 09 10:00:06 crc kubenswrapper[4824]: I1209 10:00:06.794707 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f079a0b24a09cd84dbb836f8896a1ddd826fba8e289a48331059aeb8dee3eb0b" Dec 09 10:00:06 crc kubenswrapper[4824]: I1209 10:00:06.794842 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421240-cqlx2" Dec 09 10:00:06 crc kubenswrapper[4824]: I1209 10:00:06.880041 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ca530df-dd8f-427e-bc32-6b107395b6d5-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 10:00:06 crc kubenswrapper[4824]: I1209 10:00:06.880108 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnxsr\" (UniqueName: \"kubernetes.io/projected/4ca530df-dd8f-427e-bc32-6b107395b6d5-kube-api-access-nnxsr\") on node \"crc\" DevicePath \"\"" Dec 09 10:00:06 crc kubenswrapper[4824]: I1209 10:00:06.880123 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ca530df-dd8f-427e-bc32-6b107395b6d5-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 10:00:07 crc kubenswrapper[4824]: I1209 10:00:07.805184 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"d9cee2cc-bb90-469d-8c72-29cf32fd4305","Type":"ContainerStarted","Data":"6c77615538f523e71ffb8f59dcae98178f7f96401153ef3670ed885c0b33765d"} Dec 09 10:00:07 crc kubenswrapper[4824]: I1209 10:00:07.839829 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=5.017417239 podStartE2EDuration="15.839774322s" podCreationTimestamp="2025-12-09 09:59:52 +0000 UTC" firstStartedPulling="2025-12-09 09:59:56.25589535 +0000 UTC m=+752.590400017" lastFinishedPulling="2025-12-09 10:00:07.078252433 +0000 UTC m=+763.412757100" observedRunningTime="2025-12-09 10:00:07.829727566 +0000 UTC m=+764.164232253" watchObservedRunningTime="2025-12-09 10:00:07.839774322 +0000 UTC m=+764.174278989" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.223551 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt"] Dec 09 10:00:12 crc kubenswrapper[4824]: E1209 10:00:12.225088 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ca530df-dd8f-427e-bc32-6b107395b6d5" containerName="collect-profiles" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.225123 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ca530df-dd8f-427e-bc32-6b107395b6d5" containerName="collect-profiles" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.225311 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ca530df-dd8f-427e-bc32-6b107395b6d5" containerName="collect-profiles" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.226076 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.229287 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-http" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.229596 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-config" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.229714 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-dockercfg-xw4g6" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.230134 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-grpc" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.232729 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-ca-bundle" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.238495 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt"] Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.277613 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/72ff7b94-63fd-45d1-a803-bc1ccf4388ed-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-nl2vt\" (UID: \"72ff7b94-63fd-45d1-a803-bc1ccf4388ed\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.277722 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/72ff7b94-63fd-45d1-a803-bc1ccf4388ed-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-nl2vt\" (UID: \"72ff7b94-63fd-45d1-a803-bc1ccf4388ed\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.277765 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/72ff7b94-63fd-45d1-a803-bc1ccf4388ed-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-nl2vt\" (UID: \"72ff7b94-63fd-45d1-a803-bc1ccf4388ed\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.277916 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72ff7b94-63fd-45d1-a803-bc1ccf4388ed-config\") pod \"logging-loki-distributor-76cc67bf56-nl2vt\" (UID: \"72ff7b94-63fd-45d1-a803-bc1ccf4388ed\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.277997 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl4c9\" (UniqueName: \"kubernetes.io/projected/72ff7b94-63fd-45d1-a803-bc1ccf4388ed-kube-api-access-xl4c9\") pod \"logging-loki-distributor-76cc67bf56-nl2vt\" (UID: \"72ff7b94-63fd-45d1-a803-bc1ccf4388ed\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.379738 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl4c9\" (UniqueName: \"kubernetes.io/projected/72ff7b94-63fd-45d1-a803-bc1ccf4388ed-kube-api-access-xl4c9\") pod \"logging-loki-distributor-76cc67bf56-nl2vt\" (UID: \"72ff7b94-63fd-45d1-a803-bc1ccf4388ed\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.380208 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/72ff7b94-63fd-45d1-a803-bc1ccf4388ed-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-nl2vt\" (UID: \"72ff7b94-63fd-45d1-a803-bc1ccf4388ed\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.380387 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/72ff7b94-63fd-45d1-a803-bc1ccf4388ed-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-nl2vt\" (UID: \"72ff7b94-63fd-45d1-a803-bc1ccf4388ed\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.380505 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/72ff7b94-63fd-45d1-a803-bc1ccf4388ed-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-nl2vt\" (UID: \"72ff7b94-63fd-45d1-a803-bc1ccf4388ed\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.380677 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72ff7b94-63fd-45d1-a803-bc1ccf4388ed-config\") pod \"logging-loki-distributor-76cc67bf56-nl2vt\" (UID: \"72ff7b94-63fd-45d1-a803-bc1ccf4388ed\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.381695 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/72ff7b94-63fd-45d1-a803-bc1ccf4388ed-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-nl2vt\" (UID: \"72ff7b94-63fd-45d1-a803-bc1ccf4388ed\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.382236 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72ff7b94-63fd-45d1-a803-bc1ccf4388ed-config\") pod \"logging-loki-distributor-76cc67bf56-nl2vt\" (UID: \"72ff7b94-63fd-45d1-a803-bc1ccf4388ed\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.394869 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/72ff7b94-63fd-45d1-a803-bc1ccf4388ed-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-nl2vt\" (UID: \"72ff7b94-63fd-45d1-a803-bc1ccf4388ed\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.397311 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/72ff7b94-63fd-45d1-a803-bc1ccf4388ed-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-nl2vt\" (UID: \"72ff7b94-63fd-45d1-a803-bc1ccf4388ed\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.423250 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-89dgf"] Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.424399 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.424723 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl4c9\" (UniqueName: \"kubernetes.io/projected/72ff7b94-63fd-45d1-a803-bc1ccf4388ed-kube-api-access-xl4c9\") pod \"logging-loki-distributor-76cc67bf56-nl2vt\" (UID: \"72ff7b94-63fd-45d1-a803-bc1ccf4388ed\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.427756 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-s3" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.428024 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-http" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.434342 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-grpc" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.482757 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/1006b7c9-7244-44af-8bc9-52787d891f7f-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-89dgf\" (UID: \"1006b7c9-7244-44af-8bc9-52787d891f7f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.482900 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1006b7c9-7244-44af-8bc9-52787d891f7f-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-89dgf\" (UID: \"1006b7c9-7244-44af-8bc9-52787d891f7f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.482954 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/1006b7c9-7244-44af-8bc9-52787d891f7f-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-89dgf\" (UID: \"1006b7c9-7244-44af-8bc9-52787d891f7f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.483061 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1006b7c9-7244-44af-8bc9-52787d891f7f-config\") pod \"logging-loki-querier-5895d59bb8-89dgf\" (UID: \"1006b7c9-7244-44af-8bc9-52787d891f7f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.483099 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbhd8\" (UniqueName: \"kubernetes.io/projected/1006b7c9-7244-44af-8bc9-52787d891f7f-kube-api-access-fbhd8\") pod \"logging-loki-querier-5895d59bb8-89dgf\" (UID: \"1006b7c9-7244-44af-8bc9-52787d891f7f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.483139 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/1006b7c9-7244-44af-8bc9-52787d891f7f-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-89dgf\" (UID: \"1006b7c9-7244-44af-8bc9-52787d891f7f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.511078 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-89dgf"] Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.549133 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.589450 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1006b7c9-7244-44af-8bc9-52787d891f7f-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-89dgf\" (UID: \"1006b7c9-7244-44af-8bc9-52787d891f7f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.589542 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/1006b7c9-7244-44af-8bc9-52787d891f7f-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-89dgf\" (UID: \"1006b7c9-7244-44af-8bc9-52787d891f7f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.589608 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1006b7c9-7244-44af-8bc9-52787d891f7f-config\") pod \"logging-loki-querier-5895d59bb8-89dgf\" (UID: \"1006b7c9-7244-44af-8bc9-52787d891f7f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.589633 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbhd8\" (UniqueName: \"kubernetes.io/projected/1006b7c9-7244-44af-8bc9-52787d891f7f-kube-api-access-fbhd8\") pod \"logging-loki-querier-5895d59bb8-89dgf\" (UID: \"1006b7c9-7244-44af-8bc9-52787d891f7f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.589670 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/1006b7c9-7244-44af-8bc9-52787d891f7f-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-89dgf\" (UID: \"1006b7c9-7244-44af-8bc9-52787d891f7f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.589743 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/1006b7c9-7244-44af-8bc9-52787d891f7f-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-89dgf\" (UID: \"1006b7c9-7244-44af-8bc9-52787d891f7f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.605199 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1006b7c9-7244-44af-8bc9-52787d891f7f-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-89dgf\" (UID: \"1006b7c9-7244-44af-8bc9-52787d891f7f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.606112 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/1006b7c9-7244-44af-8bc9-52787d891f7f-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-89dgf\" (UID: \"1006b7c9-7244-44af-8bc9-52787d891f7f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.607510 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1006b7c9-7244-44af-8bc9-52787d891f7f-config\") pod \"logging-loki-querier-5895d59bb8-89dgf\" (UID: \"1006b7c9-7244-44af-8bc9-52787d891f7f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.620281 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/1006b7c9-7244-44af-8bc9-52787d891f7f-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-89dgf\" (UID: \"1006b7c9-7244-44af-8bc9-52787d891f7f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.633597 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/1006b7c9-7244-44af-8bc9-52787d891f7f-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-89dgf\" (UID: \"1006b7c9-7244-44af-8bc9-52787d891f7f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.639637 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbhd8\" (UniqueName: \"kubernetes.io/projected/1006b7c9-7244-44af-8bc9-52787d891f7f-kube-api-access-fbhd8\") pod \"logging-loki-querier-5895d59bb8-89dgf\" (UID: \"1006b7c9-7244-44af-8bc9-52787d891f7f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.779211 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.789706 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm"] Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.791099 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.931998 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-http" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.932579 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-grpc" Dec 09 10:00:12 crc kubenswrapper[4824]: I1209 10:00:12.938426 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm"] Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.034851 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e05c11bd-cd92-4bac-adea-0a6049ccfb39-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-9d5fm\" (UID: \"e05c11bd-cd92-4bac-adea-0a6049ccfb39\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.034924 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/e05c11bd-cd92-4bac-adea-0a6049ccfb39-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-9d5fm\" (UID: \"e05c11bd-cd92-4bac-adea-0a6049ccfb39\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.035039 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e05c11bd-cd92-4bac-adea-0a6049ccfb39-config\") pod \"logging-loki-query-frontend-84558f7c9f-9d5fm\" (UID: \"e05c11bd-cd92-4bac-adea-0a6049ccfb39\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.035066 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4qzt\" (UniqueName: \"kubernetes.io/projected/e05c11bd-cd92-4bac-adea-0a6049ccfb39-kube-api-access-w4qzt\") pod \"logging-loki-query-frontend-84558f7c9f-9d5fm\" (UID: \"e05c11bd-cd92-4bac-adea-0a6049ccfb39\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.035117 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/e05c11bd-cd92-4bac-adea-0a6049ccfb39-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-9d5fm\" (UID: \"e05c11bd-cd92-4bac-adea-0a6049ccfb39\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.180612 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e05c11bd-cd92-4bac-adea-0a6049ccfb39-config\") pod \"logging-loki-query-frontend-84558f7c9f-9d5fm\" (UID: \"e05c11bd-cd92-4bac-adea-0a6049ccfb39\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.180661 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4qzt\" (UniqueName: \"kubernetes.io/projected/e05c11bd-cd92-4bac-adea-0a6049ccfb39-kube-api-access-w4qzt\") pod \"logging-loki-query-frontend-84558f7c9f-9d5fm\" (UID: \"e05c11bd-cd92-4bac-adea-0a6049ccfb39\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.180697 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/e05c11bd-cd92-4bac-adea-0a6049ccfb39-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-9d5fm\" (UID: \"e05c11bd-cd92-4bac-adea-0a6049ccfb39\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.180801 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e05c11bd-cd92-4bac-adea-0a6049ccfb39-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-9d5fm\" (UID: \"e05c11bd-cd92-4bac-adea-0a6049ccfb39\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.180824 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/e05c11bd-cd92-4bac-adea-0a6049ccfb39-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-9d5fm\" (UID: \"e05c11bd-cd92-4bac-adea-0a6049ccfb39\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.183711 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e05c11bd-cd92-4bac-adea-0a6049ccfb39-config\") pod \"logging-loki-query-frontend-84558f7c9f-9d5fm\" (UID: \"e05c11bd-cd92-4bac-adea-0a6049ccfb39\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.184576 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e05c11bd-cd92-4bac-adea-0a6049ccfb39-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-9d5fm\" (UID: \"e05c11bd-cd92-4bac-adea-0a6049ccfb39\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.197702 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-b549956cc-bvp6v"] Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.198625 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/e05c11bd-cd92-4bac-adea-0a6049ccfb39-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-9d5fm\" (UID: \"e05c11bd-cd92-4bac-adea-0a6049ccfb39\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.199370 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.205382 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-b549956cc-vz44w"] Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.206939 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/e05c11bd-cd92-4bac-adea-0a6049ccfb39-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-9d5fm\" (UID: \"e05c11bd-cd92-4bac-adea-0a6049ccfb39\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.207012 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.233853 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-b549956cc-vz44w"] Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.245415 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.245713 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway-ca-bundle" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.245872 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-http" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.246029 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-dockercfg-9nbpj" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.246843 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-client-http" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.247755 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.260352 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-b549956cc-bvp6v"] Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.326841 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4qzt\" (UniqueName: \"kubernetes.io/projected/e05c11bd-cd92-4bac-adea-0a6049ccfb39-kube-api-access-w4qzt\") pod \"logging-loki-query-frontend-84558f7c9f-9d5fm\" (UID: \"e05c11bd-cd92-4bac-adea-0a6049ccfb39\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.423679 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.426640 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.430030 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/93c989b8-7b86-4339-bbab-5886c7d13dc9-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-b549956cc-vz44w\" (UID: \"93c989b8-7b86-4339-bbab-5886c7d13dc9\") " pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.430101 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93c989b8-7b86-4339-bbab-5886c7d13dc9-logging-loki-ca-bundle\") pod \"logging-loki-gateway-b549956cc-vz44w\" (UID: \"93c989b8-7b86-4339-bbab-5886c7d13dc9\") " pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.430153 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/858d899b-800f-4639-8fbd-4f1ccad44991-tls-secret\") pod \"logging-loki-gateway-b549956cc-bvp6v\" (UID: \"858d899b-800f-4639-8fbd-4f1ccad44991\") " pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.430199 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/858d899b-800f-4639-8fbd-4f1ccad44991-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-b549956cc-bvp6v\" (UID: \"858d899b-800f-4639-8fbd-4f1ccad44991\") " pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.430248 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpnf9\" (UniqueName: \"kubernetes.io/projected/93c989b8-7b86-4339-bbab-5886c7d13dc9-kube-api-access-bpnf9\") pod \"logging-loki-gateway-b549956cc-vz44w\" (UID: \"93c989b8-7b86-4339-bbab-5886c7d13dc9\") " pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.430301 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/93c989b8-7b86-4339-bbab-5886c7d13dc9-tenants\") pod \"logging-loki-gateway-b549956cc-vz44w\" (UID: \"93c989b8-7b86-4339-bbab-5886c7d13dc9\") " pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.430333 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/93c989b8-7b86-4339-bbab-5886c7d13dc9-tls-secret\") pod \"logging-loki-gateway-b549956cc-vz44w\" (UID: \"93c989b8-7b86-4339-bbab-5886c7d13dc9\") " pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.430374 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/858d899b-800f-4639-8fbd-4f1ccad44991-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-b549956cc-bvp6v\" (UID: \"858d899b-800f-4639-8fbd-4f1ccad44991\") " pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.430413 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/858d899b-800f-4639-8fbd-4f1ccad44991-rbac\") pod \"logging-loki-gateway-b549956cc-bvp6v\" (UID: \"858d899b-800f-4639-8fbd-4f1ccad44991\") " pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.430444 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/93c989b8-7b86-4339-bbab-5886c7d13dc9-rbac\") pod \"logging-loki-gateway-b549956cc-vz44w\" (UID: \"93c989b8-7b86-4339-bbab-5886c7d13dc9\") " pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.430493 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/858d899b-800f-4639-8fbd-4f1ccad44991-logging-loki-ca-bundle\") pod \"logging-loki-gateway-b549956cc-bvp6v\" (UID: \"858d899b-800f-4639-8fbd-4f1ccad44991\") " pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.430532 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/93c989b8-7b86-4339-bbab-5886c7d13dc9-lokistack-gateway\") pod \"logging-loki-gateway-b549956cc-vz44w\" (UID: \"93c989b8-7b86-4339-bbab-5886c7d13dc9\") " pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.430565 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93c989b8-7b86-4339-bbab-5886c7d13dc9-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-b549956cc-vz44w\" (UID: \"93c989b8-7b86-4339-bbab-5886c7d13dc9\") " pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.430631 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkn6h\" (UniqueName: \"kubernetes.io/projected/858d899b-800f-4639-8fbd-4f1ccad44991-kube-api-access-dkn6h\") pod \"logging-loki-gateway-b549956cc-bvp6v\" (UID: \"858d899b-800f-4639-8fbd-4f1ccad44991\") " pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.430670 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/858d899b-800f-4639-8fbd-4f1ccad44991-tenants\") pod \"logging-loki-gateway-b549956cc-bvp6v\" (UID: \"858d899b-800f-4639-8fbd-4f1ccad44991\") " pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.430697 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/858d899b-800f-4639-8fbd-4f1ccad44991-lokistack-gateway\") pod \"logging-loki-gateway-b549956cc-bvp6v\" (UID: \"858d899b-800f-4639-8fbd-4f1ccad44991\") " pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.432201 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-grpc" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.432647 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-http" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.515604 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.538577 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/858d899b-800f-4639-8fbd-4f1ccad44991-logging-loki-ca-bundle\") pod \"logging-loki-gateway-b549956cc-bvp6v\" (UID: \"858d899b-800f-4639-8fbd-4f1ccad44991\") " pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.538628 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/93c989b8-7b86-4339-bbab-5886c7d13dc9-lokistack-gateway\") pod \"logging-loki-gateway-b549956cc-vz44w\" (UID: \"93c989b8-7b86-4339-bbab-5886c7d13dc9\") " pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.538654 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c50f0b79-a9b4-42df-8466-a021d5347b46\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c50f0b79-a9b4-42df-8466-a021d5347b46\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.538687 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93c989b8-7b86-4339-bbab-5886c7d13dc9-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-b549956cc-vz44w\" (UID: \"93c989b8-7b86-4339-bbab-5886c7d13dc9\") " pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.538721 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/9d24f296-729f-4eec-a1ae-d6b904399394-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.541607 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/93c989b8-7b86-4339-bbab-5886c7d13dc9-lokistack-gateway\") pod \"logging-loki-gateway-b549956cc-vz44w\" (UID: \"93c989b8-7b86-4339-bbab-5886c7d13dc9\") " pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.547540 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkn6h\" (UniqueName: \"kubernetes.io/projected/858d899b-800f-4639-8fbd-4f1ccad44991-kube-api-access-dkn6h\") pod \"logging-loki-gateway-b549956cc-bvp6v\" (UID: \"858d899b-800f-4639-8fbd-4f1ccad44991\") " pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.547596 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/9d24f296-729f-4eec-a1ae-d6b904399394-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.547646 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/9d24f296-729f-4eec-a1ae-d6b904399394-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.547668 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/858d899b-800f-4639-8fbd-4f1ccad44991-tenants\") pod \"logging-loki-gateway-b549956cc-bvp6v\" (UID: \"858d899b-800f-4639-8fbd-4f1ccad44991\") " pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.547700 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/858d899b-800f-4639-8fbd-4f1ccad44991-lokistack-gateway\") pod \"logging-loki-gateway-b549956cc-bvp6v\" (UID: \"858d899b-800f-4639-8fbd-4f1ccad44991\") " pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.547745 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/93c989b8-7b86-4339-bbab-5886c7d13dc9-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-b549956cc-vz44w\" (UID: \"93c989b8-7b86-4339-bbab-5886c7d13dc9\") " pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.547762 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93c989b8-7b86-4339-bbab-5886c7d13dc9-logging-loki-ca-bundle\") pod \"logging-loki-gateway-b549956cc-vz44w\" (UID: \"93c989b8-7b86-4339-bbab-5886c7d13dc9\") " pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.547831 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/858d899b-800f-4639-8fbd-4f1ccad44991-tls-secret\") pod \"logging-loki-gateway-b549956cc-bvp6v\" (UID: \"858d899b-800f-4639-8fbd-4f1ccad44991\") " pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.547859 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d24f296-729f-4eec-a1ae-d6b904399394-config\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.547893 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-873d4f93-375f-439e-a478-e79e760bcc44\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-873d4f93-375f-439e-a478-e79e760bcc44\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.547924 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/858d899b-800f-4639-8fbd-4f1ccad44991-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-b549956cc-bvp6v\" (UID: \"858d899b-800f-4639-8fbd-4f1ccad44991\") " pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.547965 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpnf9\" (UniqueName: \"kubernetes.io/projected/93c989b8-7b86-4339-bbab-5886c7d13dc9-kube-api-access-bpnf9\") pod \"logging-loki-gateway-b549956cc-vz44w\" (UID: \"93c989b8-7b86-4339-bbab-5886c7d13dc9\") " pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.548012 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2ngr\" (UniqueName: \"kubernetes.io/projected/9d24f296-729f-4eec-a1ae-d6b904399394-kube-api-access-f2ngr\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.548041 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/93c989b8-7b86-4339-bbab-5886c7d13dc9-tenants\") pod \"logging-loki-gateway-b549956cc-vz44w\" (UID: \"93c989b8-7b86-4339-bbab-5886c7d13dc9\") " pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.548057 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/93c989b8-7b86-4339-bbab-5886c7d13dc9-tls-secret\") pod \"logging-loki-gateway-b549956cc-vz44w\" (UID: \"93c989b8-7b86-4339-bbab-5886c7d13dc9\") " pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.548096 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/858d899b-800f-4639-8fbd-4f1ccad44991-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-b549956cc-bvp6v\" (UID: \"858d899b-800f-4639-8fbd-4f1ccad44991\") " pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.548114 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d24f296-729f-4eec-a1ae-d6b904399394-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.548141 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/858d899b-800f-4639-8fbd-4f1ccad44991-rbac\") pod \"logging-loki-gateway-b549956cc-bvp6v\" (UID: \"858d899b-800f-4639-8fbd-4f1ccad44991\") " pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.548157 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/93c989b8-7b86-4339-bbab-5886c7d13dc9-rbac\") pod \"logging-loki-gateway-b549956cc-vz44w\" (UID: \"93c989b8-7b86-4339-bbab-5886c7d13dc9\") " pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.548962 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/93c989b8-7b86-4339-bbab-5886c7d13dc9-rbac\") pod \"logging-loki-gateway-b549956cc-vz44w\" (UID: \"93c989b8-7b86-4339-bbab-5886c7d13dc9\") " pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.549544 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/858d899b-800f-4639-8fbd-4f1ccad44991-logging-loki-ca-bundle\") pod \"logging-loki-gateway-b549956cc-bvp6v\" (UID: \"858d899b-800f-4639-8fbd-4f1ccad44991\") " pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.550207 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93c989b8-7b86-4339-bbab-5886c7d13dc9-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-b549956cc-vz44w\" (UID: \"93c989b8-7b86-4339-bbab-5886c7d13dc9\") " pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.551545 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/858d899b-800f-4639-8fbd-4f1ccad44991-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-b549956cc-bvp6v\" (UID: \"858d899b-800f-4639-8fbd-4f1ccad44991\") " pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.552591 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93c989b8-7b86-4339-bbab-5886c7d13dc9-logging-loki-ca-bundle\") pod \"logging-loki-gateway-b549956cc-vz44w\" (UID: \"93c989b8-7b86-4339-bbab-5886c7d13dc9\") " pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.553541 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/858d899b-800f-4639-8fbd-4f1ccad44991-lokistack-gateway\") pod \"logging-loki-gateway-b549956cc-bvp6v\" (UID: \"858d899b-800f-4639-8fbd-4f1ccad44991\") " pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.558419 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.560551 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/858d899b-800f-4639-8fbd-4f1ccad44991-rbac\") pod \"logging-loki-gateway-b549956cc-bvp6v\" (UID: \"858d899b-800f-4639-8fbd-4f1ccad44991\") " pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.564847 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/858d899b-800f-4639-8fbd-4f1ccad44991-tls-secret\") pod \"logging-loki-gateway-b549956cc-bvp6v\" (UID: \"858d899b-800f-4639-8fbd-4f1ccad44991\") " pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.565444 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/858d899b-800f-4639-8fbd-4f1ccad44991-tenants\") pod \"logging-loki-gateway-b549956cc-bvp6v\" (UID: \"858d899b-800f-4639-8fbd-4f1ccad44991\") " pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.567918 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/93c989b8-7b86-4339-bbab-5886c7d13dc9-tls-secret\") pod \"logging-loki-gateway-b549956cc-vz44w\" (UID: \"93c989b8-7b86-4339-bbab-5886c7d13dc9\") " pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.570419 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/93c989b8-7b86-4339-bbab-5886c7d13dc9-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-b549956cc-vz44w\" (UID: \"93c989b8-7b86-4339-bbab-5886c7d13dc9\") " pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.571199 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/93c989b8-7b86-4339-bbab-5886c7d13dc9-tenants\") pod \"logging-loki-gateway-b549956cc-vz44w\" (UID: \"93c989b8-7b86-4339-bbab-5886c7d13dc9\") " pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.586056 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/858d899b-800f-4639-8fbd-4f1ccad44991-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-b549956cc-bvp6v\" (UID: \"858d899b-800f-4639-8fbd-4f1ccad44991\") " pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.601582 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpnf9\" (UniqueName: \"kubernetes.io/projected/93c989b8-7b86-4339-bbab-5886c7d13dc9-kube-api-access-bpnf9\") pod \"logging-loki-gateway-b549956cc-vz44w\" (UID: \"93c989b8-7b86-4339-bbab-5886c7d13dc9\") " pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.607026 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkn6h\" (UniqueName: \"kubernetes.io/projected/858d899b-800f-4639-8fbd-4f1ccad44991-kube-api-access-dkn6h\") pod \"logging-loki-gateway-b549956cc-bvp6v\" (UID: \"858d899b-800f-4639-8fbd-4f1ccad44991\") " pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.651622 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/9d24f296-729f-4eec-a1ae-d6b904399394-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.651965 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/9d24f296-729f-4eec-a1ae-d6b904399394-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.652119 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d24f296-729f-4eec-a1ae-d6b904399394-config\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.652233 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-873d4f93-375f-439e-a478-e79e760bcc44\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-873d4f93-375f-439e-a478-e79e760bcc44\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.652356 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2ngr\" (UniqueName: \"kubernetes.io/projected/9d24f296-729f-4eec-a1ae-d6b904399394-kube-api-access-f2ngr\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.652465 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d24f296-729f-4eec-a1ae-d6b904399394-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.652564 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c50f0b79-a9b4-42df-8466-a021d5347b46\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c50f0b79-a9b4-42df-8466-a021d5347b46\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.652677 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/9d24f296-729f-4eec-a1ae-d6b904399394-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.660225 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d24f296-729f-4eec-a1ae-d6b904399394-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.661428 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d24f296-729f-4eec-a1ae-d6b904399394-config\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.688263 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.706841 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/9d24f296-729f-4eec-a1ae-d6b904399394-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.715087 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/9d24f296-729f-4eec-a1ae-d6b904399394-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.716000 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.716036 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-873d4f93-375f-439e-a478-e79e760bcc44\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-873d4f93-375f-439e-a478-e79e760bcc44\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4f99ac9cd5ea9c850cbbfeec843e630ebbb1990dd91d27a8a9e324512e4333ea/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.716497 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.716549 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c50f0b79-a9b4-42df-8466-a021d5347b46\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c50f0b79-a9b4-42df-8466-a021d5347b46\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/af06e5560c03c301cdee1329d1b34d52dd62a6e32cb9f8cb5a754310ac0dbe21/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.717637 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/9d24f296-729f-4eec-a1ae-d6b904399394-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.724188 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.725640 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.727844 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2ngr\" (UniqueName: \"kubernetes.io/projected/9d24f296-729f-4eec-a1ae-d6b904399394-kube-api-access-f2ngr\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.743147 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-grpc" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.744687 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-http" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.755017 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.847991 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.977234 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxsj9\" (UniqueName: \"kubernetes.io/projected/379b9136-1cdf-4786-8494-e99c7161b202-kube-api-access-bxsj9\") pod \"logging-loki-compactor-0\" (UID: \"379b9136-1cdf-4786-8494-e99c7161b202\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.977339 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-09ced612-68b1-4c39-9e84-934a0ec0fe9a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-09ced612-68b1-4c39-9e84-934a0ec0fe9a\") pod \"logging-loki-compactor-0\" (UID: \"379b9136-1cdf-4786-8494-e99c7161b202\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.977428 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/379b9136-1cdf-4786-8494-e99c7161b202-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"379b9136-1cdf-4786-8494-e99c7161b202\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.977467 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/379b9136-1cdf-4786-8494-e99c7161b202-config\") pod \"logging-loki-compactor-0\" (UID: \"379b9136-1cdf-4786-8494-e99c7161b202\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.977547 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/379b9136-1cdf-4786-8494-e99c7161b202-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"379b9136-1cdf-4786-8494-e99c7161b202\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.977605 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/379b9136-1cdf-4786-8494-e99c7161b202-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"379b9136-1cdf-4786-8494-e99c7161b202\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:13 crc kubenswrapper[4824]: I1209 10:00:13.977647 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/379b9136-1cdf-4786-8494-e99c7161b202-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"379b9136-1cdf-4786-8494-e99c7161b202\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.017802 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-873d4f93-375f-439e-a478-e79e760bcc44\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-873d4f93-375f-439e-a478-e79e760bcc44\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.028683 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt"] Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.037445 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.037939 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c50f0b79-a9b4-42df-8466-a021d5347b46\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c50f0b79-a9b4-42df-8466-a021d5347b46\") pod \"logging-loki-ingester-0\" (UID: \"9d24f296-729f-4eec-a1ae-d6b904399394\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.039736 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.061470 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-grpc" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.061841 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-http" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.070040 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.085114 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/379b9136-1cdf-4786-8494-e99c7161b202-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"379b9136-1cdf-4786-8494-e99c7161b202\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.085176 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/379b9136-1cdf-4786-8494-e99c7161b202-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"379b9136-1cdf-4786-8494-e99c7161b202\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.085255 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxsj9\" (UniqueName: \"kubernetes.io/projected/379b9136-1cdf-4786-8494-e99c7161b202-kube-api-access-bxsj9\") pod \"logging-loki-compactor-0\" (UID: \"379b9136-1cdf-4786-8494-e99c7161b202\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.085297 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-09ced612-68b1-4c39-9e84-934a0ec0fe9a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-09ced612-68b1-4c39-9e84-934a0ec0fe9a\") pod \"logging-loki-compactor-0\" (UID: \"379b9136-1cdf-4786-8494-e99c7161b202\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.085345 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/379b9136-1cdf-4786-8494-e99c7161b202-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"379b9136-1cdf-4786-8494-e99c7161b202\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.085370 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/379b9136-1cdf-4786-8494-e99c7161b202-config\") pod \"logging-loki-compactor-0\" (UID: \"379b9136-1cdf-4786-8494-e99c7161b202\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.085420 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/379b9136-1cdf-4786-8494-e99c7161b202-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"379b9136-1cdf-4786-8494-e99c7161b202\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.088623 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/379b9136-1cdf-4786-8494-e99c7161b202-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"379b9136-1cdf-4786-8494-e99c7161b202\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.092172 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/379b9136-1cdf-4786-8494-e99c7161b202-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"379b9136-1cdf-4786-8494-e99c7161b202\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.094721 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/379b9136-1cdf-4786-8494-e99c7161b202-config\") pod \"logging-loki-compactor-0\" (UID: \"379b9136-1cdf-4786-8494-e99c7161b202\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.096332 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/379b9136-1cdf-4786-8494-e99c7161b202-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"379b9136-1cdf-4786-8494-e99c7161b202\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.097054 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/379b9136-1cdf-4786-8494-e99c7161b202-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"379b9136-1cdf-4786-8494-e99c7161b202\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.129514 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-89dgf"] Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.136348 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.136596 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-09ced612-68b1-4c39-9e84-934a0ec0fe9a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-09ced612-68b1-4c39-9e84-934a0ec0fe9a\") pod \"logging-loki-compactor-0\" (UID: \"379b9136-1cdf-4786-8494-e99c7161b202\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/18af7ebd7fae84a717cfe8e5481c5946d25aaa0db12fd24c22434a6324aac51c/globalmount\"" pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.148289 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxsj9\" (UniqueName: \"kubernetes.io/projected/379b9136-1cdf-4786-8494-e99c7161b202-kube-api-access-bxsj9\") pod \"logging-loki-compactor-0\" (UID: \"379b9136-1cdf-4786-8494-e99c7161b202\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.187383 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/6baf0bfe-7be6-4309-93da-e86174e4654e-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"6baf0bfe-7be6-4309-93da-e86174e4654e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.187467 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6baf0bfe-7be6-4309-93da-e86174e4654e-config\") pod \"logging-loki-index-gateway-0\" (UID: \"6baf0bfe-7be6-4309-93da-e86174e4654e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.187553 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgs77\" (UniqueName: \"kubernetes.io/projected/6baf0bfe-7be6-4309-93da-e86174e4654e-kube-api-access-mgs77\") pod \"logging-loki-index-gateway-0\" (UID: \"6baf0bfe-7be6-4309-93da-e86174e4654e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.187606 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6baf0bfe-7be6-4309-93da-e86174e4654e-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"6baf0bfe-7be6-4309-93da-e86174e4654e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.187637 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/6baf0bfe-7be6-4309-93da-e86174e4654e-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"6baf0bfe-7be6-4309-93da-e86174e4654e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.187677 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/6baf0bfe-7be6-4309-93da-e86174e4654e-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"6baf0bfe-7be6-4309-93da-e86174e4654e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.187724 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-70888067-d2b3-42e4-9404-77d866ddd376\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-70888067-d2b3-42e4-9404-77d866ddd376\") pod \"logging-loki-index-gateway-0\" (UID: \"6baf0bfe-7be6-4309-93da-e86174e4654e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.217250 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-09ced612-68b1-4c39-9e84-934a0ec0fe9a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-09ced612-68b1-4c39-9e84-934a0ec0fe9a\") pod \"logging-loki-compactor-0\" (UID: \"379b9136-1cdf-4786-8494-e99c7161b202\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.293715 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-70888067-d2b3-42e4-9404-77d866ddd376\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-70888067-d2b3-42e4-9404-77d866ddd376\") pod \"logging-loki-index-gateway-0\" (UID: \"6baf0bfe-7be6-4309-93da-e86174e4654e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.293839 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/6baf0bfe-7be6-4309-93da-e86174e4654e-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"6baf0bfe-7be6-4309-93da-e86174e4654e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.293882 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6baf0bfe-7be6-4309-93da-e86174e4654e-config\") pod \"logging-loki-index-gateway-0\" (UID: \"6baf0bfe-7be6-4309-93da-e86174e4654e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.293954 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgs77\" (UniqueName: \"kubernetes.io/projected/6baf0bfe-7be6-4309-93da-e86174e4654e-kube-api-access-mgs77\") pod \"logging-loki-index-gateway-0\" (UID: \"6baf0bfe-7be6-4309-93da-e86174e4654e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.294011 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6baf0bfe-7be6-4309-93da-e86174e4654e-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"6baf0bfe-7be6-4309-93da-e86174e4654e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.294046 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/6baf0bfe-7be6-4309-93da-e86174e4654e-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"6baf0bfe-7be6-4309-93da-e86174e4654e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.294080 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/6baf0bfe-7be6-4309-93da-e86174e4654e-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"6baf0bfe-7be6-4309-93da-e86174e4654e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.299853 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6baf0bfe-7be6-4309-93da-e86174e4654e-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"6baf0bfe-7be6-4309-93da-e86174e4654e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.300506 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6baf0bfe-7be6-4309-93da-e86174e4654e-config\") pod \"logging-loki-index-gateway-0\" (UID: \"6baf0bfe-7be6-4309-93da-e86174e4654e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.303931 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.310319 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/6baf0bfe-7be6-4309-93da-e86174e4654e-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"6baf0bfe-7be6-4309-93da-e86174e4654e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.317275 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.317434 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-70888067-d2b3-42e4-9404-77d866ddd376\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-70888067-d2b3-42e4-9404-77d866ddd376\") pod \"logging-loki-index-gateway-0\" (UID: \"6baf0bfe-7be6-4309-93da-e86174e4654e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/fde6a83d17a725fb0f9d4d1783adf866085f633de9a3a8d65443921481dce00c/globalmount\"" pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.318054 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/6baf0bfe-7be6-4309-93da-e86174e4654e-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"6baf0bfe-7be6-4309-93da-e86174e4654e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.321926 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgs77\" (UniqueName: \"kubernetes.io/projected/6baf0bfe-7be6-4309-93da-e86174e4654e-kube-api-access-mgs77\") pod \"logging-loki-index-gateway-0\" (UID: \"6baf0bfe-7be6-4309-93da-e86174e4654e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.324837 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/6baf0bfe-7be6-4309-93da-e86174e4654e-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"6baf0bfe-7be6-4309-93da-e86174e4654e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.372508 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.387898 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-70888067-d2b3-42e4-9404-77d866ddd376\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-70888067-d2b3-42e4-9404-77d866ddd376\") pod \"logging-loki-index-gateway-0\" (UID: \"6baf0bfe-7be6-4309-93da-e86174e4654e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.390658 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm"] Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.451983 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.533382 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-b549956cc-bvp6v"] Dec 09 10:00:14 crc kubenswrapper[4824]: W1209 10:00:14.559195 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod858d899b_800f_4639_8fbd_4f1ccad44991.slice/crio-b679364c6a90f7239bde5b295cecffa8e88608bcedfbe3ec34a6c207338d5023 WatchSource:0}: Error finding container b679364c6a90f7239bde5b295cecffa8e88608bcedfbe3ec34a6c207338d5023: Status 404 returned error can't find the container with id b679364c6a90f7239bde5b295cecffa8e88608bcedfbe3ec34a6c207338d5023 Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.599273 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-b549956cc-vz44w"] Dec 09 10:00:14 crc kubenswrapper[4824]: W1209 10:00:14.645554 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93c989b8_7b86_4339_bbab_5886c7d13dc9.slice/crio-2911f90f0ffc0fb87f154d0b1f4e9f81295f2fd293d7d82744b50a13a24af1a9 WatchSource:0}: Error finding container 2911f90f0ffc0fb87f154d0b1f4e9f81295f2fd293d7d82744b50a13a24af1a9: Status 404 returned error can't find the container with id 2911f90f0ffc0fb87f154d0b1f4e9f81295f2fd293d7d82744b50a13a24af1a9 Dec 09 10:00:14 crc kubenswrapper[4824]: I1209 10:00:14.893730 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 09 10:00:15 crc kubenswrapper[4824]: I1209 10:00:15.054507 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" event={"ID":"72ff7b94-63fd-45d1-a803-bc1ccf4388ed","Type":"ContainerStarted","Data":"59779ee8f7fe646dfbfc1225730aa2b2e3ce4c85fa0ffba06904f6ab682f3865"} Dec 09 10:00:15 crc kubenswrapper[4824]: I1209 10:00:15.062977 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" event={"ID":"858d899b-800f-4639-8fbd-4f1ccad44991","Type":"ContainerStarted","Data":"b679364c6a90f7239bde5b295cecffa8e88608bcedfbe3ec34a6c207338d5023"} Dec 09 10:00:15 crc kubenswrapper[4824]: I1209 10:00:15.075058 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" event={"ID":"e05c11bd-cd92-4bac-adea-0a6049ccfb39","Type":"ContainerStarted","Data":"a115eb91a7ef8f8d7516ed047fde0ac1c3f50d74b93aa2368ba0d3c820296b18"} Dec 09 10:00:15 crc kubenswrapper[4824]: I1209 10:00:15.083030 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"379b9136-1cdf-4786-8494-e99c7161b202","Type":"ContainerStarted","Data":"940df53b11c1d75cd89e8466949364a171ad9fa2a7b8bac8df3ec63630e774bb"} Dec 09 10:00:15 crc kubenswrapper[4824]: I1209 10:00:15.099825 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" event={"ID":"1006b7c9-7244-44af-8bc9-52787d891f7f","Type":"ContainerStarted","Data":"b6d55c4e71a3244907b7f53467532689ff018f65787aaba11bfaa0a9d69c91b2"} Dec 09 10:00:15 crc kubenswrapper[4824]: I1209 10:00:15.119557 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" event={"ID":"93c989b8-7b86-4339-bbab-5886c7d13dc9","Type":"ContainerStarted","Data":"2911f90f0ffc0fb87f154d0b1f4e9f81295f2fd293d7d82744b50a13a24af1a9"} Dec 09 10:00:15 crc kubenswrapper[4824]: I1209 10:00:15.132531 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 09 10:00:15 crc kubenswrapper[4824]: I1209 10:00:15.265383 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 09 10:00:15 crc kubenswrapper[4824]: W1209 10:00:15.271167 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d24f296_729f_4eec_a1ae_d6b904399394.slice/crio-8b4ea3be84570997b61eadf67ea1d1e2a6eed4c9a781e5c7d07d0288bfe3e061 WatchSource:0}: Error finding container 8b4ea3be84570997b61eadf67ea1d1e2a6eed4c9a781e5c7d07d0288bfe3e061: Status 404 returned error can't find the container with id 8b4ea3be84570997b61eadf67ea1d1e2a6eed4c9a781e5c7d07d0288bfe3e061 Dec 09 10:00:16 crc kubenswrapper[4824]: I1209 10:00:16.142040 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"9d24f296-729f-4eec-a1ae-d6b904399394","Type":"ContainerStarted","Data":"8b4ea3be84570997b61eadf67ea1d1e2a6eed4c9a781e5c7d07d0288bfe3e061"} Dec 09 10:00:16 crc kubenswrapper[4824]: I1209 10:00:16.144590 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"6baf0bfe-7be6-4309-93da-e86174e4654e","Type":"ContainerStarted","Data":"54cdcb15bb9019eedf0434df2dfd77982d399fca3b5a14c23015c707921f83ac"} Dec 09 10:00:25 crc kubenswrapper[4824]: I1209 10:00:25.832452 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"6baf0bfe-7be6-4309-93da-e86174e4654e","Type":"ContainerStarted","Data":"c4bf60e0d9cf6136a11a7bd520e1c456a0d3d36af50667373120dec32d8f1613"} Dec 09 10:00:25 crc kubenswrapper[4824]: I1209 10:00:25.833394 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:25 crc kubenswrapper[4824]: I1209 10:00:25.838829 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" event={"ID":"72ff7b94-63fd-45d1-a803-bc1ccf4388ed","Type":"ContainerStarted","Data":"996385d5e059319beba04632bd6aecd1a0ab766de567596a4d61967fd51b6630"} Dec 09 10:00:25 crc kubenswrapper[4824]: I1209 10:00:25.839083 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" Dec 09 10:00:25 crc kubenswrapper[4824]: I1209 10:00:25.841620 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" event={"ID":"858d899b-800f-4639-8fbd-4f1ccad44991","Type":"ContainerStarted","Data":"c5be1969172fe657f003b58717063cbc83d5b51701dd756e6f94b1c4f8a074c2"} Dec 09 10:00:25 crc kubenswrapper[4824]: I1209 10:00:25.845238 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" event={"ID":"e05c11bd-cd92-4bac-adea-0a6049ccfb39","Type":"ContainerStarted","Data":"bfe2c38880e90b0a96b1aa6a143863612701c31d46e36a57425d5043a8251798"} Dec 09 10:00:25 crc kubenswrapper[4824]: I1209 10:00:25.845390 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" Dec 09 10:00:25 crc kubenswrapper[4824]: I1209 10:00:25.847459 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"9d24f296-729f-4eec-a1ae-d6b904399394","Type":"ContainerStarted","Data":"390b69dd18216bf8e6503c5a24ee343ad737645dc111a226f5b2cb26a3ae7d43"} Dec 09 10:00:25 crc kubenswrapper[4824]: I1209 10:00:25.847577 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:00:25 crc kubenswrapper[4824]: I1209 10:00:25.850732 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"379b9136-1cdf-4786-8494-e99c7161b202","Type":"ContainerStarted","Data":"be01bf84bab9368277410d26d8000da80277bc7144e38fbcc26473e1e3eb06c1"} Dec 09 10:00:25 crc kubenswrapper[4824]: I1209 10:00:25.850860 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:25 crc kubenswrapper[4824]: I1209 10:00:25.854361 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" event={"ID":"1006b7c9-7244-44af-8bc9-52787d891f7f","Type":"ContainerStarted","Data":"c8c83da715ddf4e610a8dcdc87fab294a3223edc2ac63757b8314928ec56efa0"} Dec 09 10:00:25 crc kubenswrapper[4824]: I1209 10:00:25.854789 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" Dec 09 10:00:25 crc kubenswrapper[4824]: I1209 10:00:25.857662 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" event={"ID":"93c989b8-7b86-4339-bbab-5886c7d13dc9","Type":"ContainerStarted","Data":"5dd7f94bfc706c0bb190ee17e6a4b2bdefc78e5820628ef99e597f4d1af828d4"} Dec 09 10:00:26 crc kubenswrapper[4824]: I1209 10:00:26.126598 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-index-gateway-0" podStartSLOduration=4.178476393 podStartE2EDuration="14.126576956s" podCreationTimestamp="2025-12-09 10:00:12 +0000 UTC" firstStartedPulling="2025-12-09 10:00:15.154607817 +0000 UTC m=+771.489112484" lastFinishedPulling="2025-12-09 10:00:25.10270838 +0000 UTC m=+781.437213047" observedRunningTime="2025-12-09 10:00:26.119630008 +0000 UTC m=+782.454134675" watchObservedRunningTime="2025-12-09 10:00:26.126576956 +0000 UTC m=+782.461081623" Dec 09 10:00:26 crc kubenswrapper[4824]: I1209 10:00:26.174427 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" podStartSLOduration=3.442677346 podStartE2EDuration="14.174383593s" podCreationTimestamp="2025-12-09 10:00:12 +0000 UTC" firstStartedPulling="2025-12-09 10:00:14.180136627 +0000 UTC m=+770.514641294" lastFinishedPulling="2025-12-09 10:00:24.911842724 +0000 UTC m=+781.246347541" observedRunningTime="2025-12-09 10:00:26.141992153 +0000 UTC m=+782.476496810" watchObservedRunningTime="2025-12-09 10:00:26.174383593 +0000 UTC m=+782.508888260" Dec 09 10:00:26 crc kubenswrapper[4824]: I1209 10:00:26.194747 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" podStartSLOduration=3.575023967 podStartE2EDuration="14.194705484s" podCreationTimestamp="2025-12-09 10:00:12 +0000 UTC" firstStartedPulling="2025-12-09 10:00:14.396813466 +0000 UTC m=+770.731318133" lastFinishedPulling="2025-12-09 10:00:25.016494963 +0000 UTC m=+781.350999650" observedRunningTime="2025-12-09 10:00:26.173633539 +0000 UTC m=+782.508138226" watchObservedRunningTime="2025-12-09 10:00:26.194705484 +0000 UTC m=+782.529210151" Dec 09 10:00:26 crc kubenswrapper[4824]: I1209 10:00:26.213172 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-compactor-0" podStartSLOduration=4.1082211 podStartE2EDuration="14.210072608s" podCreationTimestamp="2025-12-09 10:00:12 +0000 UTC" firstStartedPulling="2025-12-09 10:00:14.914429988 +0000 UTC m=+771.248934655" lastFinishedPulling="2025-12-09 10:00:25.016281496 +0000 UTC m=+781.350786163" observedRunningTime="2025-12-09 10:00:26.201497767 +0000 UTC m=+782.536002444" watchObservedRunningTime="2025-12-09 10:00:26.210072608 +0000 UTC m=+782.544577295" Dec 09 10:00:26 crc kubenswrapper[4824]: I1209 10:00:26.238333 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" podStartSLOduration=3.248389322 podStartE2EDuration="14.237924705s" podCreationTimestamp="2025-12-09 10:00:12 +0000 UTC" firstStartedPulling="2025-12-09 10:00:14.110043698 +0000 UTC m=+770.444548365" lastFinishedPulling="2025-12-09 10:00:25.099579081 +0000 UTC m=+781.434083748" observedRunningTime="2025-12-09 10:00:26.231316708 +0000 UTC m=+782.565821395" watchObservedRunningTime="2025-12-09 10:00:26.237924705 +0000 UTC m=+782.572429372" Dec 09 10:00:26 crc kubenswrapper[4824]: I1209 10:00:26.277001 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-ingester-0" podStartSLOduration=4.42714157 podStartE2EDuration="14.276956165s" podCreationTimestamp="2025-12-09 10:00:12 +0000 UTC" firstStartedPulling="2025-12-09 10:00:15.273487534 +0000 UTC m=+771.607992201" lastFinishedPulling="2025-12-09 10:00:25.123302129 +0000 UTC m=+781.457806796" observedRunningTime="2025-12-09 10:00:26.264635398 +0000 UTC m=+782.599140085" watchObservedRunningTime="2025-12-09 10:00:26.276956165 +0000 UTC m=+782.611460842" Dec 09 10:00:30 crc kubenswrapper[4824]: I1209 10:00:30.905399 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" event={"ID":"858d899b-800f-4639-8fbd-4f1ccad44991","Type":"ContainerStarted","Data":"e71a301640493b1f819e66211ce3735dd9373f49ab72b9f2fd9437b1751139c9"} Dec 09 10:00:30 crc kubenswrapper[4824]: I1209 10:00:30.906299 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:30 crc kubenswrapper[4824]: I1209 10:00:30.906318 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:30 crc kubenswrapper[4824]: I1209 10:00:30.910066 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" event={"ID":"93c989b8-7b86-4339-bbab-5886c7d13dc9","Type":"ContainerStarted","Data":"217865726dc779ded891991c51619c6f01596838cb8161fa73701fd93c8658b0"} Dec 09 10:00:30 crc kubenswrapper[4824]: I1209 10:00:30.910644 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:30 crc kubenswrapper[4824]: I1209 10:00:30.923241 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:30 crc kubenswrapper[4824]: I1209 10:00:30.923957 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" Dec 09 10:00:30 crc kubenswrapper[4824]: I1209 10:00:30.924209 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:30 crc kubenswrapper[4824]: I1209 10:00:30.967250 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" podStartSLOduration=2.663935038 podStartE2EDuration="17.967210437s" podCreationTimestamp="2025-12-09 10:00:13 +0000 UTC" firstStartedPulling="2025-12-09 10:00:14.612187603 +0000 UTC m=+770.946692260" lastFinishedPulling="2025-12-09 10:00:29.915462982 +0000 UTC m=+786.249967659" observedRunningTime="2025-12-09 10:00:30.938051759 +0000 UTC m=+787.272556426" watchObservedRunningTime="2025-12-09 10:00:30.967210437 +0000 UTC m=+787.301715114" Dec 09 10:00:31 crc kubenswrapper[4824]: I1209 10:00:31.002178 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" podStartSLOduration=2.730626112 podStartE2EDuration="18.002143949s" podCreationTimestamp="2025-12-09 10:00:13 +0000 UTC" firstStartedPulling="2025-12-09 10:00:14.649533381 +0000 UTC m=+770.984038048" lastFinishedPulling="2025-12-09 10:00:29.921051218 +0000 UTC m=+786.255555885" observedRunningTime="2025-12-09 10:00:30.995744847 +0000 UTC m=+787.330249514" watchObservedRunningTime="2025-12-09 10:00:31.002143949 +0000 UTC m=+787.336648616" Dec 09 10:00:31 crc kubenswrapper[4824]: I1209 10:00:31.918973 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:31 crc kubenswrapper[4824]: I1209 10:00:31.931729 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" Dec 09 10:00:32 crc kubenswrapper[4824]: I1209 10:00:32.861067 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:00:32 crc kubenswrapper[4824]: I1209 10:00:32.861165 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:00:42 crc kubenswrapper[4824]: I1209 10:00:42.559235 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" Dec 09 10:00:42 crc kubenswrapper[4824]: I1209 10:00:42.788541 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" Dec 09 10:00:43 crc kubenswrapper[4824]: I1209 10:00:43.568077 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" Dec 09 10:00:44 crc kubenswrapper[4824]: I1209 10:00:44.313373 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-compactor-0" Dec 09 10:00:44 crc kubenswrapper[4824]: I1209 10:00:44.381690 4824 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Dec 09 10:00:44 crc kubenswrapper[4824]: I1209 10:00:44.381775 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="9d24f296-729f-4eec-a1ae-d6b904399394" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 09 10:00:44 crc kubenswrapper[4824]: I1209 10:00:44.463451 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 10:00:54 crc kubenswrapper[4824]: I1209 10:00:54.380191 4824 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Dec 09 10:00:54 crc kubenswrapper[4824]: I1209 10:00:54.381007 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="9d24f296-729f-4eec-a1ae-d6b904399394" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 09 10:01:02 crc kubenswrapper[4824]: I1209 10:01:02.861386 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:01:02 crc kubenswrapper[4824]: I1209 10:01:02.862071 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:01:04 crc kubenswrapper[4824]: I1209 10:01:04.379608 4824 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Dec 09 10:01:04 crc kubenswrapper[4824]: I1209 10:01:04.380143 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="9d24f296-729f-4eec-a1ae-d6b904399394" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 09 10:01:14 crc kubenswrapper[4824]: I1209 10:01:14.380386 4824 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Dec 09 10:01:14 crc kubenswrapper[4824]: I1209 10:01:14.381289 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="9d24f296-729f-4eec-a1ae-d6b904399394" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 09 10:01:24 crc kubenswrapper[4824]: I1209 10:01:24.379702 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-ingester-0" Dec 09 10:01:32 crc kubenswrapper[4824]: I1209 10:01:32.860979 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:01:32 crc kubenswrapper[4824]: I1209 10:01:32.861755 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:01:32 crc kubenswrapper[4824]: I1209 10:01:32.861843 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 10:01:32 crc kubenswrapper[4824]: I1209 10:01:32.862708 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ef93ca94e26d32583c5e2e8c602aa45c24cbdf5e3b1e4f117c890d24188f2377"} pod="openshift-machine-config-operator/machine-config-daemon-dth8x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 10:01:32 crc kubenswrapper[4824]: I1209 10:01:32.862772 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" containerID="cri-o://ef93ca94e26d32583c5e2e8c602aa45c24cbdf5e3b1e4f117c890d24188f2377" gracePeriod=600 Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.408127 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerID="ef93ca94e26d32583c5e2e8c602aa45c24cbdf5e3b1e4f117c890d24188f2377" exitCode=0 Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.408222 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerDied","Data":"ef93ca94e26d32583c5e2e8c602aa45c24cbdf5e3b1e4f117c890d24188f2377"} Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.408937 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerStarted","Data":"5abd964d470ad0e2def92ff94c5fb5265d13f56cfeabd66524120ead526f4c21"} Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.408994 4824 scope.go:117] "RemoveContainer" containerID="b9242f271427aa606711a2a58ca0863fdbf70e4a738de9f900b42d848af928bc" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.641018 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-jpm9q"] Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.642329 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.645106 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.645537 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-psm26" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.646713 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.646777 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.648469 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.653832 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.671024 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-jpm9q"] Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.713391 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/5e1fa7ff-505b-4080-98d3-2f922522cd06-metrics\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.713452 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/5e1fa7ff-505b-4080-98d3-2f922522cd06-datadir\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.713489 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/5e1fa7ff-505b-4080-98d3-2f922522cd06-sa-token\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.713514 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/5e1fa7ff-505b-4080-98d3-2f922522cd06-tmp\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.713548 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/5e1fa7ff-505b-4080-98d3-2f922522cd06-collector-token\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.713583 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh7md\" (UniqueName: \"kubernetes.io/projected/5e1fa7ff-505b-4080-98d3-2f922522cd06-kube-api-access-jh7md\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.713629 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/5e1fa7ff-505b-4080-98d3-2f922522cd06-collector-syslog-receiver\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.713665 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/5e1fa7ff-505b-4080-98d3-2f922522cd06-entrypoint\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.713713 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/5e1fa7ff-505b-4080-98d3-2f922522cd06-config-openshift-service-cacrt\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.713742 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e1fa7ff-505b-4080-98d3-2f922522cd06-trusted-ca\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.713802 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e1fa7ff-505b-4080-98d3-2f922522cd06-config\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.815032 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/5e1fa7ff-505b-4080-98d3-2f922522cd06-collector-syslog-receiver\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.815101 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/5e1fa7ff-505b-4080-98d3-2f922522cd06-entrypoint\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.815137 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/5e1fa7ff-505b-4080-98d3-2f922522cd06-config-openshift-service-cacrt\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.815165 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e1fa7ff-505b-4080-98d3-2f922522cd06-trusted-ca\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.815192 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e1fa7ff-505b-4080-98d3-2f922522cd06-config\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.815227 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/5e1fa7ff-505b-4080-98d3-2f922522cd06-metrics\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.815246 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/5e1fa7ff-505b-4080-98d3-2f922522cd06-datadir\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.815281 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/5e1fa7ff-505b-4080-98d3-2f922522cd06-sa-token\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.815302 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/5e1fa7ff-505b-4080-98d3-2f922522cd06-tmp\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.815335 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/5e1fa7ff-505b-4080-98d3-2f922522cd06-collector-token\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.815359 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh7md\" (UniqueName: \"kubernetes.io/projected/5e1fa7ff-505b-4080-98d3-2f922522cd06-kube-api-access-jh7md\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: E1209 10:01:33.815539 4824 secret.go:188] Couldn't get secret openshift-logging/collector-syslog-receiver: secret "collector-syslog-receiver" not found Dec 09 10:01:33 crc kubenswrapper[4824]: E1209 10:01:33.815613 4824 secret.go:188] Couldn't get secret openshift-logging/collector-metrics: secret "collector-metrics" not found Dec 09 10:01:33 crc kubenswrapper[4824]: E1209 10:01:33.815657 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e1fa7ff-505b-4080-98d3-2f922522cd06-collector-syslog-receiver podName:5e1fa7ff-505b-4080-98d3-2f922522cd06 nodeName:}" failed. No retries permitted until 2025-12-09 10:01:34.315628103 +0000 UTC m=+850.650132960 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "collector-syslog-receiver" (UniqueName: "kubernetes.io/secret/5e1fa7ff-505b-4080-98d3-2f922522cd06-collector-syslog-receiver") pod "collector-jpm9q" (UID: "5e1fa7ff-505b-4080-98d3-2f922522cd06") : secret "collector-syslog-receiver" not found Dec 09 10:01:33 crc kubenswrapper[4824]: E1209 10:01:33.815688 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e1fa7ff-505b-4080-98d3-2f922522cd06-metrics podName:5e1fa7ff-505b-4080-98d3-2f922522cd06 nodeName:}" failed. No retries permitted until 2025-12-09 10:01:34.315674895 +0000 UTC m=+850.650179572 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics" (UniqueName: "kubernetes.io/secret/5e1fa7ff-505b-4080-98d3-2f922522cd06-metrics") pod "collector-jpm9q" (UID: "5e1fa7ff-505b-4080-98d3-2f922522cd06") : secret "collector-metrics" not found Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.815920 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/5e1fa7ff-505b-4080-98d3-2f922522cd06-datadir\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.818168 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/5e1fa7ff-505b-4080-98d3-2f922522cd06-config-openshift-service-cacrt\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.818886 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/5e1fa7ff-505b-4080-98d3-2f922522cd06-entrypoint\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.822122 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e1fa7ff-505b-4080-98d3-2f922522cd06-trusted-ca\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.822713 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e1fa7ff-505b-4080-98d3-2f922522cd06-config\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.829366 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/5e1fa7ff-505b-4080-98d3-2f922522cd06-tmp\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.829904 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/5e1fa7ff-505b-4080-98d3-2f922522cd06-collector-token\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.833216 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-jpm9q"] Dec 09 10:01:33 crc kubenswrapper[4824]: E1209 10:01:33.834288 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[collector-syslog-receiver kube-api-access-jh7md metrics sa-token], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-logging/collector-jpm9q" podUID="5e1fa7ff-505b-4080-98d3-2f922522cd06" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.851494 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh7md\" (UniqueName: \"kubernetes.io/projected/5e1fa7ff-505b-4080-98d3-2f922522cd06-kube-api-access-jh7md\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:33 crc kubenswrapper[4824]: I1209 10:01:33.852017 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/5e1fa7ff-505b-4080-98d3-2f922522cd06-sa-token\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.323973 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/5e1fa7ff-505b-4080-98d3-2f922522cd06-metrics\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.324540 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/5e1fa7ff-505b-4080-98d3-2f922522cd06-collector-syslog-receiver\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.333816 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/5e1fa7ff-505b-4080-98d3-2f922522cd06-metrics\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.333925 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/5e1fa7ff-505b-4080-98d3-2f922522cd06-collector-syslog-receiver\") pod \"collector-jpm9q\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " pod="openshift-logging/collector-jpm9q" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.442376 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-jpm9q" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.456177 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-jpm9q" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.530934 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/5e1fa7ff-505b-4080-98d3-2f922522cd06-entrypoint\") pod \"5e1fa7ff-505b-4080-98d3-2f922522cd06\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.531044 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/5e1fa7ff-505b-4080-98d3-2f922522cd06-tmp\") pod \"5e1fa7ff-505b-4080-98d3-2f922522cd06\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.531071 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/5e1fa7ff-505b-4080-98d3-2f922522cd06-metrics\") pod \"5e1fa7ff-505b-4080-98d3-2f922522cd06\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.531190 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e1fa7ff-505b-4080-98d3-2f922522cd06-config\") pod \"5e1fa7ff-505b-4080-98d3-2f922522cd06\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.531260 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jh7md\" (UniqueName: \"kubernetes.io/projected/5e1fa7ff-505b-4080-98d3-2f922522cd06-kube-api-access-jh7md\") pod \"5e1fa7ff-505b-4080-98d3-2f922522cd06\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.531337 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/5e1fa7ff-505b-4080-98d3-2f922522cd06-datadir\") pod \"5e1fa7ff-505b-4080-98d3-2f922522cd06\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.531372 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/5e1fa7ff-505b-4080-98d3-2f922522cd06-sa-token\") pod \"5e1fa7ff-505b-4080-98d3-2f922522cd06\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.531412 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/5e1fa7ff-505b-4080-98d3-2f922522cd06-config-openshift-service-cacrt\") pod \"5e1fa7ff-505b-4080-98d3-2f922522cd06\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.531472 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/5e1fa7ff-505b-4080-98d3-2f922522cd06-collector-token\") pod \"5e1fa7ff-505b-4080-98d3-2f922522cd06\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.531507 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e1fa7ff-505b-4080-98d3-2f922522cd06-trusted-ca\") pod \"5e1fa7ff-505b-4080-98d3-2f922522cd06\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.531555 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/5e1fa7ff-505b-4080-98d3-2f922522cd06-collector-syslog-receiver\") pod \"5e1fa7ff-505b-4080-98d3-2f922522cd06\" (UID: \"5e1fa7ff-505b-4080-98d3-2f922522cd06\") " Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.531538 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e1fa7ff-505b-4080-98d3-2f922522cd06-entrypoint" (OuterVolumeSpecName: "entrypoint") pod "5e1fa7ff-505b-4080-98d3-2f922522cd06" (UID: "5e1fa7ff-505b-4080-98d3-2f922522cd06"). InnerVolumeSpecName "entrypoint". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.531584 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5e1fa7ff-505b-4080-98d3-2f922522cd06-datadir" (OuterVolumeSpecName: "datadir") pod "5e1fa7ff-505b-4080-98d3-2f922522cd06" (UID: "5e1fa7ff-505b-4080-98d3-2f922522cd06"). InnerVolumeSpecName "datadir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.532688 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e1fa7ff-505b-4080-98d3-2f922522cd06-config" (OuterVolumeSpecName: "config") pod "5e1fa7ff-505b-4080-98d3-2f922522cd06" (UID: "5e1fa7ff-505b-4080-98d3-2f922522cd06"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.533734 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e1fa7ff-505b-4080-98d3-2f922522cd06-config-openshift-service-cacrt" (OuterVolumeSpecName: "config-openshift-service-cacrt") pod "5e1fa7ff-505b-4080-98d3-2f922522cd06" (UID: "5e1fa7ff-505b-4080-98d3-2f922522cd06"). InnerVolumeSpecName "config-openshift-service-cacrt". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.534545 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e1fa7ff-505b-4080-98d3-2f922522cd06-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "5e1fa7ff-505b-4080-98d3-2f922522cd06" (UID: "5e1fa7ff-505b-4080-98d3-2f922522cd06"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.540328 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e1fa7ff-505b-4080-98d3-2f922522cd06-collector-syslog-receiver" (OuterVolumeSpecName: "collector-syslog-receiver") pod "5e1fa7ff-505b-4080-98d3-2f922522cd06" (UID: "5e1fa7ff-505b-4080-98d3-2f922522cd06"). InnerVolumeSpecName "collector-syslog-receiver". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.540342 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e1fa7ff-505b-4080-98d3-2f922522cd06-kube-api-access-jh7md" (OuterVolumeSpecName: "kube-api-access-jh7md") pod "5e1fa7ff-505b-4080-98d3-2f922522cd06" (UID: "5e1fa7ff-505b-4080-98d3-2f922522cd06"). InnerVolumeSpecName "kube-api-access-jh7md". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.540381 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e1fa7ff-505b-4080-98d3-2f922522cd06-collector-token" (OuterVolumeSpecName: "collector-token") pod "5e1fa7ff-505b-4080-98d3-2f922522cd06" (UID: "5e1fa7ff-505b-4080-98d3-2f922522cd06"). InnerVolumeSpecName "collector-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.540460 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e1fa7ff-505b-4080-98d3-2f922522cd06-tmp" (OuterVolumeSpecName: "tmp") pod "5e1fa7ff-505b-4080-98d3-2f922522cd06" (UID: "5e1fa7ff-505b-4080-98d3-2f922522cd06"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.541803 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e1fa7ff-505b-4080-98d3-2f922522cd06-sa-token" (OuterVolumeSpecName: "sa-token") pod "5e1fa7ff-505b-4080-98d3-2f922522cd06" (UID: "5e1fa7ff-505b-4080-98d3-2f922522cd06"). InnerVolumeSpecName "sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.556030 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e1fa7ff-505b-4080-98d3-2f922522cd06-metrics" (OuterVolumeSpecName: "metrics") pod "5e1fa7ff-505b-4080-98d3-2f922522cd06" (UID: "5e1fa7ff-505b-4080-98d3-2f922522cd06"). InnerVolumeSpecName "metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.633615 4824 reconciler_common.go:293] "Volume detached for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/5e1fa7ff-505b-4080-98d3-2f922522cd06-entrypoint\") on node \"crc\" DevicePath \"\"" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.633675 4824 reconciler_common.go:293] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/5e1fa7ff-505b-4080-98d3-2f922522cd06-tmp\") on node \"crc\" DevicePath \"\"" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.633693 4824 reconciler_common.go:293] "Volume detached for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/5e1fa7ff-505b-4080-98d3-2f922522cd06-metrics\") on node \"crc\" DevicePath \"\"" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.633706 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e1fa7ff-505b-4080-98d3-2f922522cd06-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.633719 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jh7md\" (UniqueName: \"kubernetes.io/projected/5e1fa7ff-505b-4080-98d3-2f922522cd06-kube-api-access-jh7md\") on node \"crc\" DevicePath \"\"" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.633735 4824 reconciler_common.go:293] "Volume detached for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/5e1fa7ff-505b-4080-98d3-2f922522cd06-datadir\") on node \"crc\" DevicePath \"\"" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.633747 4824 reconciler_common.go:293] "Volume detached for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/5e1fa7ff-505b-4080-98d3-2f922522cd06-sa-token\") on node \"crc\" DevicePath \"\"" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.633761 4824 reconciler_common.go:293] "Volume detached for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/5e1fa7ff-505b-4080-98d3-2f922522cd06-config-openshift-service-cacrt\") on node \"crc\" DevicePath \"\"" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.633774 4824 reconciler_common.go:293] "Volume detached for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/5e1fa7ff-505b-4080-98d3-2f922522cd06-collector-token\") on node \"crc\" DevicePath \"\"" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.633823 4824 reconciler_common.go:293] "Volume detached for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/5e1fa7ff-505b-4080-98d3-2f922522cd06-collector-syslog-receiver\") on node \"crc\" DevicePath \"\"" Dec 09 10:01:34 crc kubenswrapper[4824]: I1209 10:01:34.633836 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e1fa7ff-505b-4080-98d3-2f922522cd06-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.449882 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-jpm9q" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.502811 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-jpm9q"] Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.512675 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-logging/collector-jpm9q"] Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.528228 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-v9zmd"] Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.529709 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.535625 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.535686 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.536238 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-psm26" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.536518 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.536645 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.542098 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-v9zmd"] Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.548480 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.549812 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-collector-token\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.549889 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-collector-syslog-receiver\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.549922 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-config\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.549945 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-config-openshift-service-cacrt\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.549981 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-trusted-ca\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.550013 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgqvw\" (UniqueName: \"kubernetes.io/projected/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-kube-api-access-sgqvw\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.550036 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-metrics\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.550064 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-tmp\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.550088 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-entrypoint\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.550118 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-sa-token\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.550143 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-datadir\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.651679 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgqvw\" (UniqueName: \"kubernetes.io/projected/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-kube-api-access-sgqvw\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.651745 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-metrics\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.651803 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-tmp\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.651833 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-entrypoint\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.651878 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-sa-token\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.651913 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-datadir\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.651946 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-collector-token\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.651984 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-collector-syslog-receiver\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.652006 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-config\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.652032 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-config-openshift-service-cacrt\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.652071 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-trusted-ca\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.653299 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-datadir\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.654890 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-config-openshift-service-cacrt\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.654977 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-config\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.655017 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-entrypoint\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.655238 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-trusted-ca\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.656979 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-metrics\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.657914 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-collector-syslog-receiver\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.671222 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-collector-token\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.670956 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-tmp\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.674015 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgqvw\" (UniqueName: \"kubernetes.io/projected/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-kube-api-access-sgqvw\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.676216 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/91d19c8e-6e12-4dc2-a7bc-aa256d10fabe-sa-token\") pod \"collector-v9zmd\" (UID: \"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe\") " pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.854665 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-v9zmd" Dec 09 10:01:35 crc kubenswrapper[4824]: I1209 10:01:35.928851 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e1fa7ff-505b-4080-98d3-2f922522cd06" path="/var/lib/kubelet/pods/5e1fa7ff-505b-4080-98d3-2f922522cd06/volumes" Dec 09 10:01:36 crc kubenswrapper[4824]: I1209 10:01:36.295913 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-v9zmd"] Dec 09 10:01:36 crc kubenswrapper[4824]: W1209 10:01:36.304601 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91d19c8e_6e12_4dc2_a7bc_aa256d10fabe.slice/crio-0c5f21ce765e2a17669497ba71b43887c09bd9a986605e82764585c4e3bc3406 WatchSource:0}: Error finding container 0c5f21ce765e2a17669497ba71b43887c09bd9a986605e82764585c4e3bc3406: Status 404 returned error can't find the container with id 0c5f21ce765e2a17669497ba71b43887c09bd9a986605e82764585c4e3bc3406 Dec 09 10:01:36 crc kubenswrapper[4824]: I1209 10:01:36.461971 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-v9zmd" event={"ID":"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe","Type":"ContainerStarted","Data":"0c5f21ce765e2a17669497ba71b43887c09bd9a986605e82764585c4e3bc3406"} Dec 09 10:01:44 crc kubenswrapper[4824]: I1209 10:01:44.576123 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p8wdc"] Dec 09 10:01:44 crc kubenswrapper[4824]: I1209 10:01:44.579358 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p8wdc" Dec 09 10:01:44 crc kubenswrapper[4824]: I1209 10:01:44.584662 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p8wdc"] Dec 09 10:01:44 crc kubenswrapper[4824]: I1209 10:01:44.685924 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4a903dd-935d-4f79-9145-7c4bf1afe2e2-catalog-content\") pod \"redhat-operators-p8wdc\" (UID: \"f4a903dd-935d-4f79-9145-7c4bf1afe2e2\") " pod="openshift-marketplace/redhat-operators-p8wdc" Dec 09 10:01:44 crc kubenswrapper[4824]: I1209 10:01:44.685977 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4a903dd-935d-4f79-9145-7c4bf1afe2e2-utilities\") pod \"redhat-operators-p8wdc\" (UID: \"f4a903dd-935d-4f79-9145-7c4bf1afe2e2\") " pod="openshift-marketplace/redhat-operators-p8wdc" Dec 09 10:01:44 crc kubenswrapper[4824]: I1209 10:01:44.686004 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbgt8\" (UniqueName: \"kubernetes.io/projected/f4a903dd-935d-4f79-9145-7c4bf1afe2e2-kube-api-access-dbgt8\") pod \"redhat-operators-p8wdc\" (UID: \"f4a903dd-935d-4f79-9145-7c4bf1afe2e2\") " pod="openshift-marketplace/redhat-operators-p8wdc" Dec 09 10:01:44 crc kubenswrapper[4824]: I1209 10:01:44.788117 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4a903dd-935d-4f79-9145-7c4bf1afe2e2-utilities\") pod \"redhat-operators-p8wdc\" (UID: \"f4a903dd-935d-4f79-9145-7c4bf1afe2e2\") " pod="openshift-marketplace/redhat-operators-p8wdc" Dec 09 10:01:44 crc kubenswrapper[4824]: I1209 10:01:44.788208 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbgt8\" (UniqueName: \"kubernetes.io/projected/f4a903dd-935d-4f79-9145-7c4bf1afe2e2-kube-api-access-dbgt8\") pod \"redhat-operators-p8wdc\" (UID: \"f4a903dd-935d-4f79-9145-7c4bf1afe2e2\") " pod="openshift-marketplace/redhat-operators-p8wdc" Dec 09 10:01:44 crc kubenswrapper[4824]: I1209 10:01:44.788324 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4a903dd-935d-4f79-9145-7c4bf1afe2e2-catalog-content\") pod \"redhat-operators-p8wdc\" (UID: \"f4a903dd-935d-4f79-9145-7c4bf1afe2e2\") " pod="openshift-marketplace/redhat-operators-p8wdc" Dec 09 10:01:44 crc kubenswrapper[4824]: I1209 10:01:44.788935 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4a903dd-935d-4f79-9145-7c4bf1afe2e2-utilities\") pod \"redhat-operators-p8wdc\" (UID: \"f4a903dd-935d-4f79-9145-7c4bf1afe2e2\") " pod="openshift-marketplace/redhat-operators-p8wdc" Dec 09 10:01:44 crc kubenswrapper[4824]: I1209 10:01:44.791952 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4a903dd-935d-4f79-9145-7c4bf1afe2e2-catalog-content\") pod \"redhat-operators-p8wdc\" (UID: \"f4a903dd-935d-4f79-9145-7c4bf1afe2e2\") " pod="openshift-marketplace/redhat-operators-p8wdc" Dec 09 10:01:44 crc kubenswrapper[4824]: I1209 10:01:44.825051 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbgt8\" (UniqueName: \"kubernetes.io/projected/f4a903dd-935d-4f79-9145-7c4bf1afe2e2-kube-api-access-dbgt8\") pod \"redhat-operators-p8wdc\" (UID: \"f4a903dd-935d-4f79-9145-7c4bf1afe2e2\") " pod="openshift-marketplace/redhat-operators-p8wdc" Dec 09 10:01:44 crc kubenswrapper[4824]: I1209 10:01:44.903440 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p8wdc" Dec 09 10:01:49 crc kubenswrapper[4824]: I1209 10:01:49.175696 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p8wdc"] Dec 09 10:01:49 crc kubenswrapper[4824]: W1209 10:01:49.183358 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4a903dd_935d_4f79_9145_7c4bf1afe2e2.slice/crio-ff12d01fd9d6b84ead989ad3507c16198b97fcd1273036bc78970a7cebad3579 WatchSource:0}: Error finding container ff12d01fd9d6b84ead989ad3507c16198b97fcd1273036bc78970a7cebad3579: Status 404 returned error can't find the container with id ff12d01fd9d6b84ead989ad3507c16198b97fcd1273036bc78970a7cebad3579 Dec 09 10:01:49 crc kubenswrapper[4824]: I1209 10:01:49.966817 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-v9zmd" event={"ID":"91d19c8e-6e12-4dc2-a7bc-aa256d10fabe","Type":"ContainerStarted","Data":"7661c0ef1495364d0bc4bad8a5662e9fe4571fe435c5aba301588b9b71081095"} Dec 09 10:01:49 crc kubenswrapper[4824]: I1209 10:01:49.968734 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4a903dd-935d-4f79-9145-7c4bf1afe2e2" containerID="7c5afc763aedeeb31765549c730c44f2365d9a983cde9207654755a502c53a68" exitCode=0 Dec 09 10:01:49 crc kubenswrapper[4824]: I1209 10:01:49.968766 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p8wdc" event={"ID":"f4a903dd-935d-4f79-9145-7c4bf1afe2e2","Type":"ContainerDied","Data":"7c5afc763aedeeb31765549c730c44f2365d9a983cde9207654755a502c53a68"} Dec 09 10:01:49 crc kubenswrapper[4824]: I1209 10:01:49.968800 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p8wdc" event={"ID":"f4a903dd-935d-4f79-9145-7c4bf1afe2e2","Type":"ContainerStarted","Data":"ff12d01fd9d6b84ead989ad3507c16198b97fcd1273036bc78970a7cebad3579"} Dec 09 10:01:50 crc kubenswrapper[4824]: I1209 10:01:50.013494 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/collector-v9zmd" podStartSLOduration=2.337187567 podStartE2EDuration="15.01346241s" podCreationTimestamp="2025-12-09 10:01:35 +0000 UTC" firstStartedPulling="2025-12-09 10:01:36.30748994 +0000 UTC m=+852.641994607" lastFinishedPulling="2025-12-09 10:01:48.983764783 +0000 UTC m=+865.318269450" observedRunningTime="2025-12-09 10:01:49.989970529 +0000 UTC m=+866.324475226" watchObservedRunningTime="2025-12-09 10:01:50.01346241 +0000 UTC m=+866.347967087" Dec 09 10:01:52 crc kubenswrapper[4824]: I1209 10:01:52.013068 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p8wdc" event={"ID":"f4a903dd-935d-4f79-9145-7c4bf1afe2e2","Type":"ContainerStarted","Data":"811d9f21b19350a4526328481fb7c080f938c95af6540b8d67d5caa1a49c9c7e"} Dec 09 10:01:56 crc kubenswrapper[4824]: I1209 10:01:56.088131 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4a903dd-935d-4f79-9145-7c4bf1afe2e2" containerID="811d9f21b19350a4526328481fb7c080f938c95af6540b8d67d5caa1a49c9c7e" exitCode=0 Dec 09 10:01:56 crc kubenswrapper[4824]: I1209 10:01:56.088219 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p8wdc" event={"ID":"f4a903dd-935d-4f79-9145-7c4bf1afe2e2","Type":"ContainerDied","Data":"811d9f21b19350a4526328481fb7c080f938c95af6540b8d67d5caa1a49c9c7e"} Dec 09 10:01:57 crc kubenswrapper[4824]: I1209 10:01:57.100218 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p8wdc" event={"ID":"f4a903dd-935d-4f79-9145-7c4bf1afe2e2","Type":"ContainerStarted","Data":"31c094a801b316bbdd2e10c49700da2b4b960fb410db8e4f9f2e3e972fd67434"} Dec 09 10:01:57 crc kubenswrapper[4824]: I1209 10:01:57.127724 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p8wdc" podStartSLOduration=6.58450532 podStartE2EDuration="13.127701091s" podCreationTimestamp="2025-12-09 10:01:44 +0000 UTC" firstStartedPulling="2025-12-09 10:01:49.97126393 +0000 UTC m=+866.305768597" lastFinishedPulling="2025-12-09 10:01:56.514459701 +0000 UTC m=+872.848964368" observedRunningTime="2025-12-09 10:01:57.125183692 +0000 UTC m=+873.459688359" watchObservedRunningTime="2025-12-09 10:01:57.127701091 +0000 UTC m=+873.462205758" Dec 09 10:02:06 crc kubenswrapper[4824]: I1209 10:02:04.904629 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p8wdc" Dec 09 10:02:06 crc kubenswrapper[4824]: I1209 10:02:04.905536 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p8wdc" Dec 09 10:02:06 crc kubenswrapper[4824]: I1209 10:02:04.959777 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p8wdc" Dec 09 10:02:06 crc kubenswrapper[4824]: I1209 10:02:05.288161 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p8wdc" Dec 09 10:02:06 crc kubenswrapper[4824]: I1209 10:02:05.397007 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p8wdc"] Dec 09 10:02:07 crc kubenswrapper[4824]: I1209 10:02:07.191984 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p8wdc" podUID="f4a903dd-935d-4f79-9145-7c4bf1afe2e2" containerName="registry-server" containerID="cri-o://31c094a801b316bbdd2e10c49700da2b4b960fb410db8e4f9f2e3e972fd67434" gracePeriod=2 Dec 09 10:02:07 crc kubenswrapper[4824]: I1209 10:02:07.665534 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p8wdc" Dec 09 10:02:07 crc kubenswrapper[4824]: I1209 10:02:07.773315 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbgt8\" (UniqueName: \"kubernetes.io/projected/f4a903dd-935d-4f79-9145-7c4bf1afe2e2-kube-api-access-dbgt8\") pod \"f4a903dd-935d-4f79-9145-7c4bf1afe2e2\" (UID: \"f4a903dd-935d-4f79-9145-7c4bf1afe2e2\") " Dec 09 10:02:07 crc kubenswrapper[4824]: I1209 10:02:07.773437 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4a903dd-935d-4f79-9145-7c4bf1afe2e2-catalog-content\") pod \"f4a903dd-935d-4f79-9145-7c4bf1afe2e2\" (UID: \"f4a903dd-935d-4f79-9145-7c4bf1afe2e2\") " Dec 09 10:02:07 crc kubenswrapper[4824]: I1209 10:02:07.773497 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4a903dd-935d-4f79-9145-7c4bf1afe2e2-utilities\") pod \"f4a903dd-935d-4f79-9145-7c4bf1afe2e2\" (UID: \"f4a903dd-935d-4f79-9145-7c4bf1afe2e2\") " Dec 09 10:02:07 crc kubenswrapper[4824]: I1209 10:02:07.775165 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4a903dd-935d-4f79-9145-7c4bf1afe2e2-utilities" (OuterVolumeSpecName: "utilities") pod "f4a903dd-935d-4f79-9145-7c4bf1afe2e2" (UID: "f4a903dd-935d-4f79-9145-7c4bf1afe2e2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:02:07 crc kubenswrapper[4824]: I1209 10:02:07.775678 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4a903dd-935d-4f79-9145-7c4bf1afe2e2-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 10:02:07 crc kubenswrapper[4824]: I1209 10:02:07.788170 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4a903dd-935d-4f79-9145-7c4bf1afe2e2-kube-api-access-dbgt8" (OuterVolumeSpecName: "kube-api-access-dbgt8") pod "f4a903dd-935d-4f79-9145-7c4bf1afe2e2" (UID: "f4a903dd-935d-4f79-9145-7c4bf1afe2e2"). InnerVolumeSpecName "kube-api-access-dbgt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:02:07 crc kubenswrapper[4824]: I1209 10:02:07.885572 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbgt8\" (UniqueName: \"kubernetes.io/projected/f4a903dd-935d-4f79-9145-7c4bf1afe2e2-kube-api-access-dbgt8\") on node \"crc\" DevicePath \"\"" Dec 09 10:02:07 crc kubenswrapper[4824]: I1209 10:02:07.939993 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4a903dd-935d-4f79-9145-7c4bf1afe2e2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f4a903dd-935d-4f79-9145-7c4bf1afe2e2" (UID: "f4a903dd-935d-4f79-9145-7c4bf1afe2e2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:02:07 crc kubenswrapper[4824]: I1209 10:02:07.988632 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4a903dd-935d-4f79-9145-7c4bf1afe2e2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 10:02:08 crc kubenswrapper[4824]: I1209 10:02:08.204902 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4a903dd-935d-4f79-9145-7c4bf1afe2e2" containerID="31c094a801b316bbdd2e10c49700da2b4b960fb410db8e4f9f2e3e972fd67434" exitCode=0 Dec 09 10:02:08 crc kubenswrapper[4824]: I1209 10:02:08.204954 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p8wdc" event={"ID":"f4a903dd-935d-4f79-9145-7c4bf1afe2e2","Type":"ContainerDied","Data":"31c094a801b316bbdd2e10c49700da2b4b960fb410db8e4f9f2e3e972fd67434"} Dec 09 10:02:08 crc kubenswrapper[4824]: I1209 10:02:08.204992 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p8wdc" event={"ID":"f4a903dd-935d-4f79-9145-7c4bf1afe2e2","Type":"ContainerDied","Data":"ff12d01fd9d6b84ead989ad3507c16198b97fcd1273036bc78970a7cebad3579"} Dec 09 10:02:08 crc kubenswrapper[4824]: I1209 10:02:08.204990 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p8wdc" Dec 09 10:02:08 crc kubenswrapper[4824]: I1209 10:02:08.205080 4824 scope.go:117] "RemoveContainer" containerID="31c094a801b316bbdd2e10c49700da2b4b960fb410db8e4f9f2e3e972fd67434" Dec 09 10:02:08 crc kubenswrapper[4824]: I1209 10:02:08.230627 4824 scope.go:117] "RemoveContainer" containerID="811d9f21b19350a4526328481fb7c080f938c95af6540b8d67d5caa1a49c9c7e" Dec 09 10:02:08 crc kubenswrapper[4824]: I1209 10:02:08.242314 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p8wdc"] Dec 09 10:02:08 crc kubenswrapper[4824]: I1209 10:02:08.252495 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p8wdc"] Dec 09 10:02:08 crc kubenswrapper[4824]: I1209 10:02:08.262474 4824 scope.go:117] "RemoveContainer" containerID="7c5afc763aedeeb31765549c730c44f2365d9a983cde9207654755a502c53a68" Dec 09 10:02:08 crc kubenswrapper[4824]: I1209 10:02:08.281629 4824 scope.go:117] "RemoveContainer" containerID="31c094a801b316bbdd2e10c49700da2b4b960fb410db8e4f9f2e3e972fd67434" Dec 09 10:02:08 crc kubenswrapper[4824]: E1209 10:02:08.282278 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31c094a801b316bbdd2e10c49700da2b4b960fb410db8e4f9f2e3e972fd67434\": container with ID starting with 31c094a801b316bbdd2e10c49700da2b4b960fb410db8e4f9f2e3e972fd67434 not found: ID does not exist" containerID="31c094a801b316bbdd2e10c49700da2b4b960fb410db8e4f9f2e3e972fd67434" Dec 09 10:02:08 crc kubenswrapper[4824]: I1209 10:02:08.282354 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31c094a801b316bbdd2e10c49700da2b4b960fb410db8e4f9f2e3e972fd67434"} err="failed to get container status \"31c094a801b316bbdd2e10c49700da2b4b960fb410db8e4f9f2e3e972fd67434\": rpc error: code = NotFound desc = could not find container \"31c094a801b316bbdd2e10c49700da2b4b960fb410db8e4f9f2e3e972fd67434\": container with ID starting with 31c094a801b316bbdd2e10c49700da2b4b960fb410db8e4f9f2e3e972fd67434 not found: ID does not exist" Dec 09 10:02:08 crc kubenswrapper[4824]: I1209 10:02:08.282397 4824 scope.go:117] "RemoveContainer" containerID="811d9f21b19350a4526328481fb7c080f938c95af6540b8d67d5caa1a49c9c7e" Dec 09 10:02:08 crc kubenswrapper[4824]: E1209 10:02:08.293175 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"811d9f21b19350a4526328481fb7c080f938c95af6540b8d67d5caa1a49c9c7e\": container with ID starting with 811d9f21b19350a4526328481fb7c080f938c95af6540b8d67d5caa1a49c9c7e not found: ID does not exist" containerID="811d9f21b19350a4526328481fb7c080f938c95af6540b8d67d5caa1a49c9c7e" Dec 09 10:02:08 crc kubenswrapper[4824]: I1209 10:02:08.293234 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"811d9f21b19350a4526328481fb7c080f938c95af6540b8d67d5caa1a49c9c7e"} err="failed to get container status \"811d9f21b19350a4526328481fb7c080f938c95af6540b8d67d5caa1a49c9c7e\": rpc error: code = NotFound desc = could not find container \"811d9f21b19350a4526328481fb7c080f938c95af6540b8d67d5caa1a49c9c7e\": container with ID starting with 811d9f21b19350a4526328481fb7c080f938c95af6540b8d67d5caa1a49c9c7e not found: ID does not exist" Dec 09 10:02:08 crc kubenswrapper[4824]: I1209 10:02:08.293277 4824 scope.go:117] "RemoveContainer" containerID="7c5afc763aedeeb31765549c730c44f2365d9a983cde9207654755a502c53a68" Dec 09 10:02:08 crc kubenswrapper[4824]: E1209 10:02:08.293761 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c5afc763aedeeb31765549c730c44f2365d9a983cde9207654755a502c53a68\": container with ID starting with 7c5afc763aedeeb31765549c730c44f2365d9a983cde9207654755a502c53a68 not found: ID does not exist" containerID="7c5afc763aedeeb31765549c730c44f2365d9a983cde9207654755a502c53a68" Dec 09 10:02:08 crc kubenswrapper[4824]: I1209 10:02:08.293853 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c5afc763aedeeb31765549c730c44f2365d9a983cde9207654755a502c53a68"} err="failed to get container status \"7c5afc763aedeeb31765549c730c44f2365d9a983cde9207654755a502c53a68\": rpc error: code = NotFound desc = could not find container \"7c5afc763aedeeb31765549c730c44f2365d9a983cde9207654755a502c53a68\": container with ID starting with 7c5afc763aedeeb31765549c730c44f2365d9a983cde9207654755a502c53a68 not found: ID does not exist" Dec 09 10:02:09 crc kubenswrapper[4824]: I1209 10:02:09.919149 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4a903dd-935d-4f79-9145-7c4bf1afe2e2" path="/var/lib/kubelet/pods/f4a903dd-935d-4f79-9145-7c4bf1afe2e2/volumes" Dec 09 10:02:20 crc kubenswrapper[4824]: I1209 10:02:20.489738 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc"] Dec 09 10:02:20 crc kubenswrapper[4824]: E1209 10:02:20.490759 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4a903dd-935d-4f79-9145-7c4bf1afe2e2" containerName="extract-content" Dec 09 10:02:20 crc kubenswrapper[4824]: I1209 10:02:20.490795 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4a903dd-935d-4f79-9145-7c4bf1afe2e2" containerName="extract-content" Dec 09 10:02:20 crc kubenswrapper[4824]: E1209 10:02:20.490815 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4a903dd-935d-4f79-9145-7c4bf1afe2e2" containerName="extract-utilities" Dec 09 10:02:20 crc kubenswrapper[4824]: I1209 10:02:20.490824 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4a903dd-935d-4f79-9145-7c4bf1afe2e2" containerName="extract-utilities" Dec 09 10:02:20 crc kubenswrapper[4824]: E1209 10:02:20.490841 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4a903dd-935d-4f79-9145-7c4bf1afe2e2" containerName="registry-server" Dec 09 10:02:20 crc kubenswrapper[4824]: I1209 10:02:20.490850 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4a903dd-935d-4f79-9145-7c4bf1afe2e2" containerName="registry-server" Dec 09 10:02:20 crc kubenswrapper[4824]: I1209 10:02:20.491024 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4a903dd-935d-4f79-9145-7c4bf1afe2e2" containerName="registry-server" Dec 09 10:02:20 crc kubenswrapper[4824]: I1209 10:02:20.492415 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc" Dec 09 10:02:20 crc kubenswrapper[4824]: I1209 10:02:20.495374 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 09 10:02:20 crc kubenswrapper[4824]: I1209 10:02:20.502510 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc"] Dec 09 10:02:20 crc kubenswrapper[4824]: I1209 10:02:20.560311 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/42a87103-eb40-4e22-97d7-18ff2885d50b-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc\" (UID: \"42a87103-eb40-4e22-97d7-18ff2885d50b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc" Dec 09 10:02:20 crc kubenswrapper[4824]: I1209 10:02:20.560406 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/42a87103-eb40-4e22-97d7-18ff2885d50b-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc\" (UID: \"42a87103-eb40-4e22-97d7-18ff2885d50b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc" Dec 09 10:02:20 crc kubenswrapper[4824]: I1209 10:02:20.560438 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5wbw\" (UniqueName: \"kubernetes.io/projected/42a87103-eb40-4e22-97d7-18ff2885d50b-kube-api-access-z5wbw\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc\" (UID: \"42a87103-eb40-4e22-97d7-18ff2885d50b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc" Dec 09 10:02:20 crc kubenswrapper[4824]: I1209 10:02:20.661756 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/42a87103-eb40-4e22-97d7-18ff2885d50b-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc\" (UID: \"42a87103-eb40-4e22-97d7-18ff2885d50b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc" Dec 09 10:02:20 crc kubenswrapper[4824]: I1209 10:02:20.661878 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/42a87103-eb40-4e22-97d7-18ff2885d50b-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc\" (UID: \"42a87103-eb40-4e22-97d7-18ff2885d50b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc" Dec 09 10:02:20 crc kubenswrapper[4824]: I1209 10:02:20.661919 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5wbw\" (UniqueName: \"kubernetes.io/projected/42a87103-eb40-4e22-97d7-18ff2885d50b-kube-api-access-z5wbw\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc\" (UID: \"42a87103-eb40-4e22-97d7-18ff2885d50b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc" Dec 09 10:02:20 crc kubenswrapper[4824]: I1209 10:02:20.662511 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/42a87103-eb40-4e22-97d7-18ff2885d50b-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc\" (UID: \"42a87103-eb40-4e22-97d7-18ff2885d50b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc" Dec 09 10:02:20 crc kubenswrapper[4824]: I1209 10:02:20.662823 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/42a87103-eb40-4e22-97d7-18ff2885d50b-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc\" (UID: \"42a87103-eb40-4e22-97d7-18ff2885d50b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc" Dec 09 10:02:20 crc kubenswrapper[4824]: I1209 10:02:20.684865 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5wbw\" (UniqueName: \"kubernetes.io/projected/42a87103-eb40-4e22-97d7-18ff2885d50b-kube-api-access-z5wbw\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc\" (UID: \"42a87103-eb40-4e22-97d7-18ff2885d50b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc" Dec 09 10:02:20 crc kubenswrapper[4824]: I1209 10:02:20.818348 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc" Dec 09 10:02:21 crc kubenswrapper[4824]: I1209 10:02:21.274316 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc"] Dec 09 10:02:21 crc kubenswrapper[4824]: I1209 10:02:21.329813 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc" event={"ID":"42a87103-eb40-4e22-97d7-18ff2885d50b","Type":"ContainerStarted","Data":"2fd8f365409dae019e9d63484fec978fc65ced7c0ffad7101720d66d2c5cc339"} Dec 09 10:02:22 crc kubenswrapper[4824]: I1209 10:02:22.340388 4824 generic.go:334] "Generic (PLEG): container finished" podID="42a87103-eb40-4e22-97d7-18ff2885d50b" containerID="715aac5048e9ebd9f366c4055bdd1aeb16518d343ef6824c1949f68a8d4bf653" exitCode=0 Dec 09 10:02:22 crc kubenswrapper[4824]: I1209 10:02:22.340738 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc" event={"ID":"42a87103-eb40-4e22-97d7-18ff2885d50b","Type":"ContainerDied","Data":"715aac5048e9ebd9f366c4055bdd1aeb16518d343ef6824c1949f68a8d4bf653"} Dec 09 10:02:22 crc kubenswrapper[4824]: I1209 10:02:22.342801 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 10:02:24 crc kubenswrapper[4824]: I1209 10:02:24.357275 4824 generic.go:334] "Generic (PLEG): container finished" podID="42a87103-eb40-4e22-97d7-18ff2885d50b" containerID="083434973b5e73abec80e5a49cafd1f2e6161eb53b5cfd1e7ae227f57cc59b9b" exitCode=0 Dec 09 10:02:24 crc kubenswrapper[4824]: I1209 10:02:24.357512 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc" event={"ID":"42a87103-eb40-4e22-97d7-18ff2885d50b","Type":"ContainerDied","Data":"083434973b5e73abec80e5a49cafd1f2e6161eb53b5cfd1e7ae227f57cc59b9b"} Dec 09 10:02:25 crc kubenswrapper[4824]: I1209 10:02:25.367575 4824 generic.go:334] "Generic (PLEG): container finished" podID="42a87103-eb40-4e22-97d7-18ff2885d50b" containerID="e4285e78ab50bef286a4dffb1924cf47d86763b1b74ebf7b425a87d2e1754332" exitCode=0 Dec 09 10:02:25 crc kubenswrapper[4824]: I1209 10:02:25.367635 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc" event={"ID":"42a87103-eb40-4e22-97d7-18ff2885d50b","Type":"ContainerDied","Data":"e4285e78ab50bef286a4dffb1924cf47d86763b1b74ebf7b425a87d2e1754332"} Dec 09 10:02:26 crc kubenswrapper[4824]: I1209 10:02:26.654149 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc" Dec 09 10:02:26 crc kubenswrapper[4824]: I1209 10:02:26.702568 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/42a87103-eb40-4e22-97d7-18ff2885d50b-bundle\") pod \"42a87103-eb40-4e22-97d7-18ff2885d50b\" (UID: \"42a87103-eb40-4e22-97d7-18ff2885d50b\") " Dec 09 10:02:26 crc kubenswrapper[4824]: I1209 10:02:26.702661 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5wbw\" (UniqueName: \"kubernetes.io/projected/42a87103-eb40-4e22-97d7-18ff2885d50b-kube-api-access-z5wbw\") pod \"42a87103-eb40-4e22-97d7-18ff2885d50b\" (UID: \"42a87103-eb40-4e22-97d7-18ff2885d50b\") " Dec 09 10:02:26 crc kubenswrapper[4824]: I1209 10:02:26.702819 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/42a87103-eb40-4e22-97d7-18ff2885d50b-util\") pod \"42a87103-eb40-4e22-97d7-18ff2885d50b\" (UID: \"42a87103-eb40-4e22-97d7-18ff2885d50b\") " Dec 09 10:02:26 crc kubenswrapper[4824]: I1209 10:02:26.703665 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42a87103-eb40-4e22-97d7-18ff2885d50b-bundle" (OuterVolumeSpecName: "bundle") pod "42a87103-eb40-4e22-97d7-18ff2885d50b" (UID: "42a87103-eb40-4e22-97d7-18ff2885d50b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:02:26 crc kubenswrapper[4824]: I1209 10:02:26.710223 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42a87103-eb40-4e22-97d7-18ff2885d50b-kube-api-access-z5wbw" (OuterVolumeSpecName: "kube-api-access-z5wbw") pod "42a87103-eb40-4e22-97d7-18ff2885d50b" (UID: "42a87103-eb40-4e22-97d7-18ff2885d50b"). InnerVolumeSpecName "kube-api-access-z5wbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:02:26 crc kubenswrapper[4824]: I1209 10:02:26.718410 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42a87103-eb40-4e22-97d7-18ff2885d50b-util" (OuterVolumeSpecName: "util") pod "42a87103-eb40-4e22-97d7-18ff2885d50b" (UID: "42a87103-eb40-4e22-97d7-18ff2885d50b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:02:26 crc kubenswrapper[4824]: I1209 10:02:26.805316 4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/42a87103-eb40-4e22-97d7-18ff2885d50b-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:02:26 crc kubenswrapper[4824]: I1209 10:02:26.805366 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5wbw\" (UniqueName: \"kubernetes.io/projected/42a87103-eb40-4e22-97d7-18ff2885d50b-kube-api-access-z5wbw\") on node \"crc\" DevicePath \"\"" Dec 09 10:02:26 crc kubenswrapper[4824]: I1209 10:02:26.805380 4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/42a87103-eb40-4e22-97d7-18ff2885d50b-util\") on node \"crc\" DevicePath \"\"" Dec 09 10:02:27 crc kubenswrapper[4824]: I1209 10:02:27.385511 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc" event={"ID":"42a87103-eb40-4e22-97d7-18ff2885d50b","Type":"ContainerDied","Data":"2fd8f365409dae019e9d63484fec978fc65ced7c0ffad7101720d66d2c5cc339"} Dec 09 10:02:27 crc kubenswrapper[4824]: I1209 10:02:27.385567 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc" Dec 09 10:02:27 crc kubenswrapper[4824]: I1209 10:02:27.385569 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fd8f365409dae019e9d63484fec978fc65ced7c0ffad7101720d66d2c5cc339" Dec 09 10:02:30 crc kubenswrapper[4824]: I1209 10:02:30.222887 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-45dxs"] Dec 09 10:02:30 crc kubenswrapper[4824]: E1209 10:02:30.223592 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42a87103-eb40-4e22-97d7-18ff2885d50b" containerName="extract" Dec 09 10:02:30 crc kubenswrapper[4824]: I1209 10:02:30.223610 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="42a87103-eb40-4e22-97d7-18ff2885d50b" containerName="extract" Dec 09 10:02:30 crc kubenswrapper[4824]: E1209 10:02:30.223631 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42a87103-eb40-4e22-97d7-18ff2885d50b" containerName="pull" Dec 09 10:02:30 crc kubenswrapper[4824]: I1209 10:02:30.223639 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="42a87103-eb40-4e22-97d7-18ff2885d50b" containerName="pull" Dec 09 10:02:30 crc kubenswrapper[4824]: E1209 10:02:30.223662 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42a87103-eb40-4e22-97d7-18ff2885d50b" containerName="util" Dec 09 10:02:30 crc kubenswrapper[4824]: I1209 10:02:30.223670 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="42a87103-eb40-4e22-97d7-18ff2885d50b" containerName="util" Dec 09 10:02:30 crc kubenswrapper[4824]: I1209 10:02:30.223876 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="42a87103-eb40-4e22-97d7-18ff2885d50b" containerName="extract" Dec 09 10:02:30 crc kubenswrapper[4824]: I1209 10:02:30.224556 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-45dxs" Dec 09 10:02:30 crc kubenswrapper[4824]: I1209 10:02:30.226689 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-qqnmz" Dec 09 10:02:30 crc kubenswrapper[4824]: I1209 10:02:30.227887 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 09 10:02:30 crc kubenswrapper[4824]: I1209 10:02:30.228206 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 09 10:02:30 crc kubenswrapper[4824]: I1209 10:02:30.234457 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-45dxs"] Dec 09 10:02:30 crc kubenswrapper[4824]: I1209 10:02:30.261969 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68d6c\" (UniqueName: \"kubernetes.io/projected/d9389a26-39b2-43f3-bd7b-54f83b8060ed-kube-api-access-68d6c\") pod \"nmstate-operator-5b5b58f5c8-45dxs\" (UID: \"d9389a26-39b2-43f3-bd7b-54f83b8060ed\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-45dxs" Dec 09 10:02:30 crc kubenswrapper[4824]: I1209 10:02:30.363527 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68d6c\" (UniqueName: \"kubernetes.io/projected/d9389a26-39b2-43f3-bd7b-54f83b8060ed-kube-api-access-68d6c\") pod \"nmstate-operator-5b5b58f5c8-45dxs\" (UID: \"d9389a26-39b2-43f3-bd7b-54f83b8060ed\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-45dxs" Dec 09 10:02:30 crc kubenswrapper[4824]: I1209 10:02:30.384218 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68d6c\" (UniqueName: \"kubernetes.io/projected/d9389a26-39b2-43f3-bd7b-54f83b8060ed-kube-api-access-68d6c\") pod \"nmstate-operator-5b5b58f5c8-45dxs\" (UID: \"d9389a26-39b2-43f3-bd7b-54f83b8060ed\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-45dxs" Dec 09 10:02:30 crc kubenswrapper[4824]: I1209 10:02:30.544864 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-45dxs" Dec 09 10:02:31 crc kubenswrapper[4824]: I1209 10:02:31.049847 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-45dxs"] Dec 09 10:02:31 crc kubenswrapper[4824]: I1209 10:02:31.413245 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-45dxs" event={"ID":"d9389a26-39b2-43f3-bd7b-54f83b8060ed","Type":"ContainerStarted","Data":"fabcc8a49fa91166e9aff2d30f99d86a2e7d56674500f4891137052146495d81"} Dec 09 10:02:31 crc kubenswrapper[4824]: I1209 10:02:31.598929 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-w2ml5"] Dec 09 10:02:31 crc kubenswrapper[4824]: I1209 10:02:31.600490 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w2ml5" Dec 09 10:02:31 crc kubenswrapper[4824]: I1209 10:02:31.609074 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w2ml5"] Dec 09 10:02:31 crc kubenswrapper[4824]: I1209 10:02:31.693120 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290-catalog-content\") pod \"community-operators-w2ml5\" (UID: \"6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290\") " pod="openshift-marketplace/community-operators-w2ml5" Dec 09 10:02:31 crc kubenswrapper[4824]: I1209 10:02:31.693360 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l98lh\" (UniqueName: \"kubernetes.io/projected/6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290-kube-api-access-l98lh\") pod \"community-operators-w2ml5\" (UID: \"6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290\") " pod="openshift-marketplace/community-operators-w2ml5" Dec 09 10:02:31 crc kubenswrapper[4824]: I1209 10:02:31.693455 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290-utilities\") pod \"community-operators-w2ml5\" (UID: \"6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290\") " pod="openshift-marketplace/community-operators-w2ml5" Dec 09 10:02:31 crc kubenswrapper[4824]: I1209 10:02:31.795035 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l98lh\" (UniqueName: \"kubernetes.io/projected/6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290-kube-api-access-l98lh\") pod \"community-operators-w2ml5\" (UID: \"6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290\") " pod="openshift-marketplace/community-operators-w2ml5" Dec 09 10:02:31 crc kubenswrapper[4824]: I1209 10:02:31.795136 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290-utilities\") pod \"community-operators-w2ml5\" (UID: \"6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290\") " pod="openshift-marketplace/community-operators-w2ml5" Dec 09 10:02:31 crc kubenswrapper[4824]: I1209 10:02:31.795202 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290-catalog-content\") pod \"community-operators-w2ml5\" (UID: \"6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290\") " pod="openshift-marketplace/community-operators-w2ml5" Dec 09 10:02:31 crc kubenswrapper[4824]: I1209 10:02:31.796024 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290-catalog-content\") pod \"community-operators-w2ml5\" (UID: \"6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290\") " pod="openshift-marketplace/community-operators-w2ml5" Dec 09 10:02:31 crc kubenswrapper[4824]: I1209 10:02:31.796144 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290-utilities\") pod \"community-operators-w2ml5\" (UID: \"6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290\") " pod="openshift-marketplace/community-operators-w2ml5" Dec 09 10:02:31 crc kubenswrapper[4824]: I1209 10:02:31.818757 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l98lh\" (UniqueName: \"kubernetes.io/projected/6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290-kube-api-access-l98lh\") pod \"community-operators-w2ml5\" (UID: \"6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290\") " pod="openshift-marketplace/community-operators-w2ml5" Dec 09 10:02:31 crc kubenswrapper[4824]: I1209 10:02:31.927454 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w2ml5" Dec 09 10:02:32 crc kubenswrapper[4824]: I1209 10:02:32.351294 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w2ml5"] Dec 09 10:02:32 crc kubenswrapper[4824]: I1209 10:02:32.428977 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2ml5" event={"ID":"6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290","Type":"ContainerStarted","Data":"3c2de135c8782842578df13c9cf9042e4c871a2c004a06e69793399a83ec4723"} Dec 09 10:02:33 crc kubenswrapper[4824]: I1209 10:02:33.438414 4824 generic.go:334] "Generic (PLEG): container finished" podID="6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290" containerID="ec3eb415a4b0efbb88342e1a7ceef71ed80b36ec10109071e47545d5520ce0eb" exitCode=0 Dec 09 10:02:33 crc kubenswrapper[4824]: I1209 10:02:33.438489 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2ml5" event={"ID":"6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290","Type":"ContainerDied","Data":"ec3eb415a4b0efbb88342e1a7ceef71ed80b36ec10109071e47545d5520ce0eb"} Dec 09 10:02:34 crc kubenswrapper[4824]: I1209 10:02:34.450588 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2ml5" event={"ID":"6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290","Type":"ContainerStarted","Data":"fcf2bbf9ff10f2ae16ce8656521f85024524c16ad6767314bb48e4de01fa3acb"} Dec 09 10:02:34 crc kubenswrapper[4824]: E1209 10:02:34.618627 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d49df1d_a9d9_4e2a_bfec_4b0dc4be0290.slice/crio-fcf2bbf9ff10f2ae16ce8656521f85024524c16ad6767314bb48e4de01fa3acb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d49df1d_a9d9_4e2a_bfec_4b0dc4be0290.slice/crio-conmon-fcf2bbf9ff10f2ae16ce8656521f85024524c16ad6767314bb48e4de01fa3acb.scope\": RecentStats: unable to find data in memory cache]" Dec 09 10:02:35 crc kubenswrapper[4824]: I1209 10:02:35.472516 4824 generic.go:334] "Generic (PLEG): container finished" podID="6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290" containerID="fcf2bbf9ff10f2ae16ce8656521f85024524c16ad6767314bb48e4de01fa3acb" exitCode=0 Dec 09 10:02:35 crc kubenswrapper[4824]: I1209 10:02:35.472630 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2ml5" event={"ID":"6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290","Type":"ContainerDied","Data":"fcf2bbf9ff10f2ae16ce8656521f85024524c16ad6767314bb48e4de01fa3acb"} Dec 09 10:02:36 crc kubenswrapper[4824]: I1209 10:02:36.485040 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2ml5" event={"ID":"6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290","Type":"ContainerStarted","Data":"c6bbcb043e205ed3ce52040f278e2f624028890cf1c59d97b0d7add41594c2ec"} Dec 09 10:02:36 crc kubenswrapper[4824]: I1209 10:02:36.509401 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-w2ml5" podStartSLOduration=2.952582524 podStartE2EDuration="5.509381209s" podCreationTimestamp="2025-12-09 10:02:31 +0000 UTC" firstStartedPulling="2025-12-09 10:02:33.440443554 +0000 UTC m=+909.774948221" lastFinishedPulling="2025-12-09 10:02:35.997242239 +0000 UTC m=+912.331746906" observedRunningTime="2025-12-09 10:02:36.507727456 +0000 UTC m=+912.842232133" watchObservedRunningTime="2025-12-09 10:02:36.509381209 +0000 UTC m=+912.843885876" Dec 09 10:02:39 crc kubenswrapper[4824]: I1209 10:02:39.518863 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-45dxs" event={"ID":"d9389a26-39b2-43f3-bd7b-54f83b8060ed","Type":"ContainerStarted","Data":"22999c61de26898960332880b796a1a6fde3ad882c158e6a99b9bad375409bcc"} Dec 09 10:02:39 crc kubenswrapper[4824]: I1209 10:02:39.541508 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-45dxs" podStartSLOduration=1.8662180689999999 podStartE2EDuration="9.541487834s" podCreationTimestamp="2025-12-09 10:02:30 +0000 UTC" firstStartedPulling="2025-12-09 10:02:31.059008968 +0000 UTC m=+907.393513655" lastFinishedPulling="2025-12-09 10:02:38.734278753 +0000 UTC m=+915.068783420" observedRunningTime="2025-12-09 10:02:39.538172969 +0000 UTC m=+915.872677626" watchObservedRunningTime="2025-12-09 10:02:39.541487834 +0000 UTC m=+915.875992501" Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.208218 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-44svc"] Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.210120 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-44svc" Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.227028 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-44svc"] Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.368849 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40b723c1-2e4c-4f13-948c-8db2b4aad424-catalog-content\") pod \"redhat-marketplace-44svc\" (UID: \"40b723c1-2e4c-4f13-948c-8db2b4aad424\") " pod="openshift-marketplace/redhat-marketplace-44svc" Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.369005 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40b723c1-2e4c-4f13-948c-8db2b4aad424-utilities\") pod \"redhat-marketplace-44svc\" (UID: \"40b723c1-2e4c-4f13-948c-8db2b4aad424\") " pod="openshift-marketplace/redhat-marketplace-44svc" Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.369064 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fjcl\" (UniqueName: \"kubernetes.io/projected/40b723c1-2e4c-4f13-948c-8db2b4aad424-kube-api-access-4fjcl\") pod \"redhat-marketplace-44svc\" (UID: \"40b723c1-2e4c-4f13-948c-8db2b4aad424\") " pod="openshift-marketplace/redhat-marketplace-44svc" Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.470520 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40b723c1-2e4c-4f13-948c-8db2b4aad424-catalog-content\") pod \"redhat-marketplace-44svc\" (UID: \"40b723c1-2e4c-4f13-948c-8db2b4aad424\") " pod="openshift-marketplace/redhat-marketplace-44svc" Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.470678 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40b723c1-2e4c-4f13-948c-8db2b4aad424-utilities\") pod \"redhat-marketplace-44svc\" (UID: \"40b723c1-2e4c-4f13-948c-8db2b4aad424\") " pod="openshift-marketplace/redhat-marketplace-44svc" Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.470746 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fjcl\" (UniqueName: \"kubernetes.io/projected/40b723c1-2e4c-4f13-948c-8db2b4aad424-kube-api-access-4fjcl\") pod \"redhat-marketplace-44svc\" (UID: \"40b723c1-2e4c-4f13-948c-8db2b4aad424\") " pod="openshift-marketplace/redhat-marketplace-44svc" Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.471448 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40b723c1-2e4c-4f13-948c-8db2b4aad424-catalog-content\") pod \"redhat-marketplace-44svc\" (UID: \"40b723c1-2e4c-4f13-948c-8db2b4aad424\") " pod="openshift-marketplace/redhat-marketplace-44svc" Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.471514 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40b723c1-2e4c-4f13-948c-8db2b4aad424-utilities\") pod \"redhat-marketplace-44svc\" (UID: \"40b723c1-2e4c-4f13-948c-8db2b4aad424\") " pod="openshift-marketplace/redhat-marketplace-44svc" Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.517459 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fjcl\" (UniqueName: \"kubernetes.io/projected/40b723c1-2e4c-4f13-948c-8db2b4aad424-kube-api-access-4fjcl\") pod \"redhat-marketplace-44svc\" (UID: \"40b723c1-2e4c-4f13-948c-8db2b4aad424\") " pod="openshift-marketplace/redhat-marketplace-44svc" Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.531712 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-44svc" Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.862774 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-z8zrc"] Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.874487 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-z8zrc" Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.879302 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-z8zrc"] Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.883349 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgtcd\" (UniqueName: \"kubernetes.io/projected/1927c56b-adc3-4088-8d46-8cb3cf9acc1c-kube-api-access-sgtcd\") pod \"nmstate-metrics-7f946cbc9-z8zrc\" (UID: \"1927c56b-adc3-4088-8d46-8cb3cf9acc1c\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-z8zrc" Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.886484 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-jch2q" Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.899725 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jklpr"] Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.901556 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jklpr" Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.906720 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.922038 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jklpr"] Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.933257 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-5strj"] Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.934582 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-5strj" Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.997831 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/952171d0-1a2e-4801-bc11-4012ba19588c-dbus-socket\") pod \"nmstate-handler-5strj\" (UID: \"952171d0-1a2e-4801-bc11-4012ba19588c\") " pod="openshift-nmstate/nmstate-handler-5strj" Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.998017 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/952171d0-1a2e-4801-bc11-4012ba19588c-ovs-socket\") pod \"nmstate-handler-5strj\" (UID: \"952171d0-1a2e-4801-bc11-4012ba19588c\") " pod="openshift-nmstate/nmstate-handler-5strj" Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.998070 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgtcd\" (UniqueName: \"kubernetes.io/projected/1927c56b-adc3-4088-8d46-8cb3cf9acc1c-kube-api-access-sgtcd\") pod \"nmstate-metrics-7f946cbc9-z8zrc\" (UID: \"1927c56b-adc3-4088-8d46-8cb3cf9acc1c\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-z8zrc" Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.998115 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/952171d0-1a2e-4801-bc11-4012ba19588c-nmstate-lock\") pod \"nmstate-handler-5strj\" (UID: \"952171d0-1a2e-4801-bc11-4012ba19588c\") " pod="openshift-nmstate/nmstate-handler-5strj" Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.998294 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d68qt\" (UniqueName: \"kubernetes.io/projected/952171d0-1a2e-4801-bc11-4012ba19588c-kube-api-access-d68qt\") pod \"nmstate-handler-5strj\" (UID: \"952171d0-1a2e-4801-bc11-4012ba19588c\") " pod="openshift-nmstate/nmstate-handler-5strj" Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.998443 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/0b932274-5eba-4e0d-8a64-1f469a6ab3d1-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-jklpr\" (UID: \"0b932274-5eba-4e0d-8a64-1f469a6ab3d1\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jklpr" Dec 09 10:02:40 crc kubenswrapper[4824]: I1209 10:02:40.998499 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9x42b\" (UniqueName: \"kubernetes.io/projected/0b932274-5eba-4e0d-8a64-1f469a6ab3d1-kube-api-access-9x42b\") pod \"nmstate-webhook-5f6d4c5ccb-jklpr\" (UID: \"0b932274-5eba-4e0d-8a64-1f469a6ab3d1\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jklpr" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.047256 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgtcd\" (UniqueName: \"kubernetes.io/projected/1927c56b-adc3-4088-8d46-8cb3cf9acc1c-kube-api-access-sgtcd\") pod \"nmstate-metrics-7f946cbc9-z8zrc\" (UID: \"1927c56b-adc3-4088-8d46-8cb3cf9acc1c\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-z8zrc" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.103325 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d68qt\" (UniqueName: \"kubernetes.io/projected/952171d0-1a2e-4801-bc11-4012ba19588c-kube-api-access-d68qt\") pod \"nmstate-handler-5strj\" (UID: \"952171d0-1a2e-4801-bc11-4012ba19588c\") " pod="openshift-nmstate/nmstate-handler-5strj" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.103422 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/0b932274-5eba-4e0d-8a64-1f469a6ab3d1-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-jklpr\" (UID: \"0b932274-5eba-4e0d-8a64-1f469a6ab3d1\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jklpr" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.103452 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9x42b\" (UniqueName: \"kubernetes.io/projected/0b932274-5eba-4e0d-8a64-1f469a6ab3d1-kube-api-access-9x42b\") pod \"nmstate-webhook-5f6d4c5ccb-jklpr\" (UID: \"0b932274-5eba-4e0d-8a64-1f469a6ab3d1\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jklpr" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.103483 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/952171d0-1a2e-4801-bc11-4012ba19588c-dbus-socket\") pod \"nmstate-handler-5strj\" (UID: \"952171d0-1a2e-4801-bc11-4012ba19588c\") " pod="openshift-nmstate/nmstate-handler-5strj" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.103516 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/952171d0-1a2e-4801-bc11-4012ba19588c-ovs-socket\") pod \"nmstate-handler-5strj\" (UID: \"952171d0-1a2e-4801-bc11-4012ba19588c\") " pod="openshift-nmstate/nmstate-handler-5strj" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.103540 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/952171d0-1a2e-4801-bc11-4012ba19588c-nmstate-lock\") pod \"nmstate-handler-5strj\" (UID: \"952171d0-1a2e-4801-bc11-4012ba19588c\") " pod="openshift-nmstate/nmstate-handler-5strj" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.103614 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/952171d0-1a2e-4801-bc11-4012ba19588c-nmstate-lock\") pod \"nmstate-handler-5strj\" (UID: \"952171d0-1a2e-4801-bc11-4012ba19588c\") " pod="openshift-nmstate/nmstate-handler-5strj" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.104380 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/952171d0-1a2e-4801-bc11-4012ba19588c-ovs-socket\") pod \"nmstate-handler-5strj\" (UID: \"952171d0-1a2e-4801-bc11-4012ba19588c\") " pod="openshift-nmstate/nmstate-handler-5strj" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.104645 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/952171d0-1a2e-4801-bc11-4012ba19588c-dbus-socket\") pod \"nmstate-handler-5strj\" (UID: \"952171d0-1a2e-4801-bc11-4012ba19588c\") " pod="openshift-nmstate/nmstate-handler-5strj" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.110557 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/0b932274-5eba-4e0d-8a64-1f469a6ab3d1-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-jklpr\" (UID: \"0b932274-5eba-4e0d-8a64-1f469a6ab3d1\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jklpr" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.112952 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-w6n8z"] Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.114119 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-w6n8z" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.117292 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.122317 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-xsg9q" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.122591 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.138563 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-w6n8z"] Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.147652 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d68qt\" (UniqueName: \"kubernetes.io/projected/952171d0-1a2e-4801-bc11-4012ba19588c-kube-api-access-d68qt\") pod \"nmstate-handler-5strj\" (UID: \"952171d0-1a2e-4801-bc11-4012ba19588c\") " pod="openshift-nmstate/nmstate-handler-5strj" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.171700 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9x42b\" (UniqueName: \"kubernetes.io/projected/0b932274-5eba-4e0d-8a64-1f469a6ab3d1-kube-api-access-9x42b\") pod \"nmstate-webhook-5f6d4c5ccb-jklpr\" (UID: \"0b932274-5eba-4e0d-8a64-1f469a6ab3d1\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jklpr" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.213594 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-44svc"] Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.218725 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-z8zrc" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.231164 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jklpr" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.256907 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-5strj" Dec 09 10:02:41 crc kubenswrapper[4824]: W1209 10:02:41.275523 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40b723c1_2e4c_4f13_948c_8db2b4aad424.slice/crio-3436f0a8ee0b59a457c8e515c565579c2c5dcc71408f58375ae6e7cc42c43ddd WatchSource:0}: Error finding container 3436f0a8ee0b59a457c8e515c565579c2c5dcc71408f58375ae6e7cc42c43ddd: Status 404 returned error can't find the container with id 3436f0a8ee0b59a457c8e515c565579c2c5dcc71408f58375ae6e7cc42c43ddd Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.307618 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th2bv\" (UniqueName: \"kubernetes.io/projected/d9469d6d-b886-4086-8c80-9ea36f65574d-kube-api-access-th2bv\") pod \"nmstate-console-plugin-7fbb5f6569-w6n8z\" (UID: \"d9469d6d-b886-4086-8c80-9ea36f65574d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-w6n8z" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.307675 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d9469d6d-b886-4086-8c80-9ea36f65574d-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-w6n8z\" (UID: \"d9469d6d-b886-4086-8c80-9ea36f65574d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-w6n8z" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.307704 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d9469d6d-b886-4086-8c80-9ea36f65574d-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-w6n8z\" (UID: \"d9469d6d-b886-4086-8c80-9ea36f65574d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-w6n8z" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.413021 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th2bv\" (UniqueName: \"kubernetes.io/projected/d9469d6d-b886-4086-8c80-9ea36f65574d-kube-api-access-th2bv\") pod \"nmstate-console-plugin-7fbb5f6569-w6n8z\" (UID: \"d9469d6d-b886-4086-8c80-9ea36f65574d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-w6n8z" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.413111 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d9469d6d-b886-4086-8c80-9ea36f65574d-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-w6n8z\" (UID: \"d9469d6d-b886-4086-8c80-9ea36f65574d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-w6n8z" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.413200 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d9469d6d-b886-4086-8c80-9ea36f65574d-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-w6n8z\" (UID: \"d9469d6d-b886-4086-8c80-9ea36f65574d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-w6n8z" Dec 09 10:02:41 crc kubenswrapper[4824]: E1209 10:02:41.413806 4824 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 09 10:02:41 crc kubenswrapper[4824]: E1209 10:02:41.413878 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d9469d6d-b886-4086-8c80-9ea36f65574d-plugin-serving-cert podName:d9469d6d-b886-4086-8c80-9ea36f65574d nodeName:}" failed. No retries permitted until 2025-12-09 10:02:41.913852364 +0000 UTC m=+918.248357031 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/d9469d6d-b886-4086-8c80-9ea36f65574d-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-w6n8z" (UID: "d9469d6d-b886-4086-8c80-9ea36f65574d") : secret "plugin-serving-cert" not found Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.414757 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-77dcf56988-s8mzf"] Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.415036 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d9469d6d-b886-4086-8c80-9ea36f65574d-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-w6n8z\" (UID: \"d9469d6d-b886-4086-8c80-9ea36f65574d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-w6n8z" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.420224 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.439518 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th2bv\" (UniqueName: \"kubernetes.io/projected/d9469d6d-b886-4086-8c80-9ea36f65574d-kube-api-access-th2bv\") pod \"nmstate-console-plugin-7fbb5f6569-w6n8z\" (UID: \"d9469d6d-b886-4086-8c80-9ea36f65574d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-w6n8z" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.474196 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-77dcf56988-s8mzf"] Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.569015 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-5strj" event={"ID":"952171d0-1a2e-4801-bc11-4012ba19588c","Type":"ContainerStarted","Data":"02e17bb319cb7bbd9c360d28a4634055163cac5833da628b19d361f4d9645b95"} Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.576888 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-44svc" event={"ID":"40b723c1-2e4c-4f13-948c-8db2b4aad424","Type":"ContainerStarted","Data":"3436f0a8ee0b59a457c8e515c565579c2c5dcc71408f58375ae6e7cc42c43ddd"} Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.617736 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/721e1031-d80c-406a-ad29-17879d58c3c3-console-config\") pod \"console-77dcf56988-s8mzf\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.617837 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/721e1031-d80c-406a-ad29-17879d58c3c3-console-serving-cert\") pod \"console-77dcf56988-s8mzf\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.617997 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj46v\" (UniqueName: \"kubernetes.io/projected/721e1031-d80c-406a-ad29-17879d58c3c3-kube-api-access-gj46v\") pod \"console-77dcf56988-s8mzf\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.618061 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/721e1031-d80c-406a-ad29-17879d58c3c3-service-ca\") pod \"console-77dcf56988-s8mzf\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.618206 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/721e1031-d80c-406a-ad29-17879d58c3c3-console-oauth-config\") pod \"console-77dcf56988-s8mzf\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.618323 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/721e1031-d80c-406a-ad29-17879d58c3c3-trusted-ca-bundle\") pod \"console-77dcf56988-s8mzf\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.618387 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/721e1031-d80c-406a-ad29-17879d58c3c3-oauth-serving-cert\") pod \"console-77dcf56988-s8mzf\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.722842 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/721e1031-d80c-406a-ad29-17879d58c3c3-console-config\") pod \"console-77dcf56988-s8mzf\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.723480 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/721e1031-d80c-406a-ad29-17879d58c3c3-console-serving-cert\") pod \"console-77dcf56988-s8mzf\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.723555 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj46v\" (UniqueName: \"kubernetes.io/projected/721e1031-d80c-406a-ad29-17879d58c3c3-kube-api-access-gj46v\") pod \"console-77dcf56988-s8mzf\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.723595 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/721e1031-d80c-406a-ad29-17879d58c3c3-service-ca\") pod \"console-77dcf56988-s8mzf\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.723725 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/721e1031-d80c-406a-ad29-17879d58c3c3-console-oauth-config\") pod \"console-77dcf56988-s8mzf\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.723822 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/721e1031-d80c-406a-ad29-17879d58c3c3-trusted-ca-bundle\") pod \"console-77dcf56988-s8mzf\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.723982 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/721e1031-d80c-406a-ad29-17879d58c3c3-oauth-serving-cert\") pod \"console-77dcf56988-s8mzf\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.725028 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/721e1031-d80c-406a-ad29-17879d58c3c3-console-config\") pod \"console-77dcf56988-s8mzf\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.725641 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/721e1031-d80c-406a-ad29-17879d58c3c3-service-ca\") pod \"console-77dcf56988-s8mzf\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.726225 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/721e1031-d80c-406a-ad29-17879d58c3c3-trusted-ca-bundle\") pod \"console-77dcf56988-s8mzf\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.727138 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/721e1031-d80c-406a-ad29-17879d58c3c3-oauth-serving-cert\") pod \"console-77dcf56988-s8mzf\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.747343 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/721e1031-d80c-406a-ad29-17879d58c3c3-console-oauth-config\") pod \"console-77dcf56988-s8mzf\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.761314 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/721e1031-d80c-406a-ad29-17879d58c3c3-console-serving-cert\") pod \"console-77dcf56988-s8mzf\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.766752 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj46v\" (UniqueName: \"kubernetes.io/projected/721e1031-d80c-406a-ad29-17879d58c3c3-kube-api-access-gj46v\") pod \"console-77dcf56988-s8mzf\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.806078 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.928370 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d9469d6d-b886-4086-8c80-9ea36f65574d-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-w6n8z\" (UID: \"d9469d6d-b886-4086-8c80-9ea36f65574d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-w6n8z" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.932217 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d9469d6d-b886-4086-8c80-9ea36f65574d-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-w6n8z\" (UID: \"d9469d6d-b886-4086-8c80-9ea36f65574d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-w6n8z" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.950285 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-w2ml5" Dec 09 10:02:41 crc kubenswrapper[4824]: I1209 10:02:41.950330 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-w2ml5" Dec 09 10:02:42 crc kubenswrapper[4824]: I1209 10:02:42.000379 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-w2ml5" Dec 09 10:02:42 crc kubenswrapper[4824]: I1209 10:02:42.079945 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-w6n8z" Dec 09 10:02:42 crc kubenswrapper[4824]: I1209 10:02:42.145429 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-z8zrc"] Dec 09 10:02:42 crc kubenswrapper[4824]: W1209 10:02:42.152038 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1927c56b_adc3_4088_8d46_8cb3cf9acc1c.slice/crio-fd7aad33d3682f7ac23319b161652d217314056c55d294c07089d5fbe762a4db WatchSource:0}: Error finding container fd7aad33d3682f7ac23319b161652d217314056c55d294c07089d5fbe762a4db: Status 404 returned error can't find the container with id fd7aad33d3682f7ac23319b161652d217314056c55d294c07089d5fbe762a4db Dec 09 10:02:42 crc kubenswrapper[4824]: I1209 10:02:42.249768 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jklpr"] Dec 09 10:02:42 crc kubenswrapper[4824]: I1209 10:02:42.442915 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-77dcf56988-s8mzf"] Dec 09 10:02:42 crc kubenswrapper[4824]: I1209 10:02:42.592838 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-z8zrc" event={"ID":"1927c56b-adc3-4088-8d46-8cb3cf9acc1c","Type":"ContainerStarted","Data":"fd7aad33d3682f7ac23319b161652d217314056c55d294c07089d5fbe762a4db"} Dec 09 10:02:42 crc kubenswrapper[4824]: I1209 10:02:42.595276 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-77dcf56988-s8mzf" event={"ID":"721e1031-d80c-406a-ad29-17879d58c3c3","Type":"ContainerStarted","Data":"f31650000a80adda0dbae2781d24e63e5ca5da99fa3ba17b6a2588f41cca02cd"} Dec 09 10:02:42 crc kubenswrapper[4824]: I1209 10:02:42.597148 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jklpr" event={"ID":"0b932274-5eba-4e0d-8a64-1f469a6ab3d1","Type":"ContainerStarted","Data":"01c98b3dd36eccfa0496ba17637c0b4703b8746eeccde55f89a268c67d51c32d"} Dec 09 10:02:42 crc kubenswrapper[4824]: I1209 10:02:42.599221 4824 generic.go:334] "Generic (PLEG): container finished" podID="40b723c1-2e4c-4f13-948c-8db2b4aad424" containerID="84492123746f740e5ed1f8b21dd84ae821743aa0ea56915950f319a16d70e6e0" exitCode=0 Dec 09 10:02:42 crc kubenswrapper[4824]: I1209 10:02:42.599485 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-44svc" event={"ID":"40b723c1-2e4c-4f13-948c-8db2b4aad424","Type":"ContainerDied","Data":"84492123746f740e5ed1f8b21dd84ae821743aa0ea56915950f319a16d70e6e0"} Dec 09 10:02:42 crc kubenswrapper[4824]: I1209 10:02:42.660399 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-w2ml5" Dec 09 10:02:42 crc kubenswrapper[4824]: I1209 10:02:42.673013 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-w6n8z"] Dec 09 10:02:43 crc kubenswrapper[4824]: I1209 10:02:43.623572 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-w6n8z" event={"ID":"d9469d6d-b886-4086-8c80-9ea36f65574d","Type":"ContainerStarted","Data":"bcfc68fb1a0f2a5b1cf6ef17398016111de770ab6f0103d9a891724c6caa1ee3"} Dec 09 10:02:43 crc kubenswrapper[4824]: I1209 10:02:43.639796 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-77dcf56988-s8mzf" event={"ID":"721e1031-d80c-406a-ad29-17879d58c3c3","Type":"ContainerStarted","Data":"88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857"} Dec 09 10:02:43 crc kubenswrapper[4824]: I1209 10:02:43.679418 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-77dcf56988-s8mzf" podStartSLOduration=2.67939405 podStartE2EDuration="2.67939405s" podCreationTimestamp="2025-12-09 10:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:02:43.673342229 +0000 UTC m=+920.007846906" watchObservedRunningTime="2025-12-09 10:02:43.67939405 +0000 UTC m=+920.013898717" Dec 09 10:02:44 crc kubenswrapper[4824]: I1209 10:02:44.405451 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w2ml5"] Dec 09 10:02:44 crc kubenswrapper[4824]: I1209 10:02:44.666840 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-44svc" event={"ID":"40b723c1-2e4c-4f13-948c-8db2b4aad424","Type":"ContainerStarted","Data":"f82221ae45609757673eb7d09537d232b45d154559bfc59d1e90e862735dc5e1"} Dec 09 10:02:44 crc kubenswrapper[4824]: I1209 10:02:44.667977 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-w2ml5" podUID="6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290" containerName="registry-server" containerID="cri-o://c6bbcb043e205ed3ce52040f278e2f624028890cf1c59d97b0d7add41594c2ec" gracePeriod=2 Dec 09 10:02:44 crc kubenswrapper[4824]: E1209 10:02:44.858963 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d49df1d_a9d9_4e2a_bfec_4b0dc4be0290.slice/crio-c6bbcb043e205ed3ce52040f278e2f624028890cf1c59d97b0d7add41594c2ec.scope\": RecentStats: unable to find data in memory cache]" Dec 09 10:02:45 crc kubenswrapper[4824]: I1209 10:02:45.725231 4824 generic.go:334] "Generic (PLEG): container finished" podID="40b723c1-2e4c-4f13-948c-8db2b4aad424" containerID="f82221ae45609757673eb7d09537d232b45d154559bfc59d1e90e862735dc5e1" exitCode=0 Dec 09 10:02:45 crc kubenswrapper[4824]: I1209 10:02:45.725871 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-44svc" event={"ID":"40b723c1-2e4c-4f13-948c-8db2b4aad424","Type":"ContainerDied","Data":"f82221ae45609757673eb7d09537d232b45d154559bfc59d1e90e862735dc5e1"} Dec 09 10:02:45 crc kubenswrapper[4824]: I1209 10:02:45.754047 4824 generic.go:334] "Generic (PLEG): container finished" podID="6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290" containerID="c6bbcb043e205ed3ce52040f278e2f624028890cf1c59d97b0d7add41594c2ec" exitCode=0 Dec 09 10:02:45 crc kubenswrapper[4824]: I1209 10:02:45.754121 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2ml5" event={"ID":"6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290","Type":"ContainerDied","Data":"c6bbcb043e205ed3ce52040f278e2f624028890cf1c59d97b0d7add41594c2ec"} Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.106365 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w2ml5" Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.146669 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290-utilities\") pod \"6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290\" (UID: \"6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290\") " Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.147118 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290-catalog-content\") pod \"6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290\" (UID: \"6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290\") " Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.147153 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l98lh\" (UniqueName: \"kubernetes.io/projected/6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290-kube-api-access-l98lh\") pod \"6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290\" (UID: \"6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290\") " Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.149543 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290-utilities" (OuterVolumeSpecName: "utilities") pod "6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290" (UID: "6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.154559 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290-kube-api-access-l98lh" (OuterVolumeSpecName: "kube-api-access-l98lh") pod "6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290" (UID: "6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290"). InnerVolumeSpecName "kube-api-access-l98lh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.200601 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290" (UID: "6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.251132 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.251166 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.251177 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l98lh\" (UniqueName: \"kubernetes.io/projected/6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290-kube-api-access-l98lh\") on node \"crc\" DevicePath \"\"" Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.788540 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2ml5" event={"ID":"6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290","Type":"ContainerDied","Data":"3c2de135c8782842578df13c9cf9042e4c871a2c004a06e69793399a83ec4723"} Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.789068 4824 scope.go:117] "RemoveContainer" containerID="c6bbcb043e205ed3ce52040f278e2f624028890cf1c59d97b0d7add41594c2ec" Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.788575 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w2ml5" Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.790998 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-w6n8z" event={"ID":"d9469d6d-b886-4086-8c80-9ea36f65574d","Type":"ContainerStarted","Data":"732586c067db394554cc629f911dde9d30854d83b30e2a111108e5f4efb7eaee"} Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.799800 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-5strj" event={"ID":"952171d0-1a2e-4801-bc11-4012ba19588c","Type":"ContainerStarted","Data":"03b015e008bee436f1cc7928ceee5f96d62c16f5b1af877e06a372d8eeedb66e"} Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.809922 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-z8zrc" event={"ID":"1927c56b-adc3-4088-8d46-8cb3cf9acc1c","Type":"ContainerStarted","Data":"dbf6a3f4d72496e9c795114b5f36f6221ebaf7610c15d8c0860f481fb841ce77"} Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.811749 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jklpr" event={"ID":"0b932274-5eba-4e0d-8a64-1f469a6ab3d1","Type":"ContainerStarted","Data":"ae3899f3649704cb6da3480d63068571693f579072a50a6416fb682413b5e41e"} Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.812989 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jklpr" Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.838571 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-44svc" event={"ID":"40b723c1-2e4c-4f13-948c-8db2b4aad424","Type":"ContainerStarted","Data":"fd4b3bf21aead16cfe1de302b14bf0e9443790b047c135b46a16adefd149b26b"} Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.845426 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w2ml5"] Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.860204 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-w2ml5"] Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.877921 4824 scope.go:117] "RemoveContainer" containerID="fcf2bbf9ff10f2ae16ce8656521f85024524c16ad6767314bb48e4de01fa3acb" Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.888864 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-44svc" podStartSLOduration=3.327398394 podStartE2EDuration="8.888839397s" podCreationTimestamp="2025-12-09 10:02:40 +0000 UTC" firstStartedPulling="2025-12-09 10:02:42.603566994 +0000 UTC m=+918.938071661" lastFinishedPulling="2025-12-09 10:02:48.165007997 +0000 UTC m=+924.499512664" observedRunningTime="2025-12-09 10:02:48.878615334 +0000 UTC m=+925.213120011" watchObservedRunningTime="2025-12-09 10:02:48.888839397 +0000 UTC m=+925.223344064" Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.918444 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jklpr" podStartSLOduration=3.032432906 podStartE2EDuration="8.91840449s" podCreationTimestamp="2025-12-09 10:02:40 +0000 UTC" firstStartedPulling="2025-12-09 10:02:42.280416147 +0000 UTC m=+918.614920814" lastFinishedPulling="2025-12-09 10:02:48.166387731 +0000 UTC m=+924.500892398" observedRunningTime="2025-12-09 10:02:48.904584204 +0000 UTC m=+925.239088871" watchObservedRunningTime="2025-12-09 10:02:48.91840449 +0000 UTC m=+925.252909157" Dec 09 10:02:48 crc kubenswrapper[4824]: I1209 10:02:48.929850 4824 scope.go:117] "RemoveContainer" containerID="ec3eb415a4b0efbb88342e1a7ceef71ed80b36ec10109071e47545d5520ce0eb" Dec 09 10:02:49 crc kubenswrapper[4824]: I1209 10:02:49.849843 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-5strj" Dec 09 10:02:49 crc kubenswrapper[4824]: I1209 10:02:49.868480 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-5strj" podStartSLOduration=3.240201111 podStartE2EDuration="9.868460377s" podCreationTimestamp="2025-12-09 10:02:40 +0000 UTC" firstStartedPulling="2025-12-09 10:02:41.540156019 +0000 UTC m=+917.874660676" lastFinishedPulling="2025-12-09 10:02:48.168415275 +0000 UTC m=+924.502919942" observedRunningTime="2025-12-09 10:02:49.867425394 +0000 UTC m=+926.201930081" watchObservedRunningTime="2025-12-09 10:02:49.868460377 +0000 UTC m=+926.202965044" Dec 09 10:02:49 crc kubenswrapper[4824]: I1209 10:02:49.891906 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-w6n8z" podStartSLOduration=3.425996458 podStartE2EDuration="8.891878077s" podCreationTimestamp="2025-12-09 10:02:41 +0000 UTC" firstStartedPulling="2025-12-09 10:02:42.70138174 +0000 UTC m=+919.035886407" lastFinishedPulling="2025-12-09 10:02:48.167263369 +0000 UTC m=+924.501768026" observedRunningTime="2025-12-09 10:02:49.8894679 +0000 UTC m=+926.223972587" watchObservedRunningTime="2025-12-09 10:02:49.891878077 +0000 UTC m=+926.226382754" Dec 09 10:02:49 crc kubenswrapper[4824]: I1209 10:02:49.924180 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290" path="/var/lib/kubelet/pods/6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290/volumes" Dec 09 10:02:50 crc kubenswrapper[4824]: I1209 10:02:50.532139 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-44svc" Dec 09 10:02:50 crc kubenswrapper[4824]: I1209 10:02:50.532399 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-44svc" Dec 09 10:02:51 crc kubenswrapper[4824]: I1209 10:02:51.591140 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-44svc" podUID="40b723c1-2e4c-4f13-948c-8db2b4aad424" containerName="registry-server" probeResult="failure" output=< Dec 09 10:02:51 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 10:02:51 crc kubenswrapper[4824]: > Dec 09 10:02:51 crc kubenswrapper[4824]: I1209 10:02:51.807802 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:51 crc kubenswrapper[4824]: I1209 10:02:51.807854 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:51 crc kubenswrapper[4824]: I1209 10:02:51.814278 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:51 crc kubenswrapper[4824]: I1209 10:02:51.878842 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:02:51 crc kubenswrapper[4824]: I1209 10:02:51.950766 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-79d58c8b9c-dbzmt"] Dec 09 10:02:52 crc kubenswrapper[4824]: I1209 10:02:52.894225 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-z8zrc" event={"ID":"1927c56b-adc3-4088-8d46-8cb3cf9acc1c","Type":"ContainerStarted","Data":"0d27416f6ce692f1e001d812295026abfd171d89fc5089607a21eb344f4a42cf"} Dec 09 10:02:52 crc kubenswrapper[4824]: I1209 10:02:52.930940 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-z8zrc" podStartSLOduration=2.422450119 podStartE2EDuration="12.930912769s" podCreationTimestamp="2025-12-09 10:02:40 +0000 UTC" firstStartedPulling="2025-12-09 10:02:42.169615601 +0000 UTC m=+918.504120268" lastFinishedPulling="2025-12-09 10:02:52.678078251 +0000 UTC m=+929.012582918" observedRunningTime="2025-12-09 10:02:52.915146222 +0000 UTC m=+929.249650899" watchObservedRunningTime="2025-12-09 10:02:52.930912769 +0000 UTC m=+929.265417436" Dec 09 10:02:56 crc kubenswrapper[4824]: I1209 10:02:56.289743 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-5strj" Dec 09 10:03:00 crc kubenswrapper[4824]: I1209 10:03:00.587034 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-44svc" Dec 09 10:03:00 crc kubenswrapper[4824]: I1209 10:03:00.632609 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-44svc" Dec 09 10:03:00 crc kubenswrapper[4824]: I1209 10:03:00.824027 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-44svc"] Dec 09 10:03:01 crc kubenswrapper[4824]: I1209 10:03:01.237406 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jklpr" Dec 09 10:03:01 crc kubenswrapper[4824]: I1209 10:03:01.981729 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-44svc" podUID="40b723c1-2e4c-4f13-948c-8db2b4aad424" containerName="registry-server" containerID="cri-o://fd4b3bf21aead16cfe1de302b14bf0e9443790b047c135b46a16adefd149b26b" gracePeriod=2 Dec 09 10:03:02 crc kubenswrapper[4824]: I1209 10:03:02.507363 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-44svc" Dec 09 10:03:02 crc kubenswrapper[4824]: I1209 10:03:02.626803 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fjcl\" (UniqueName: \"kubernetes.io/projected/40b723c1-2e4c-4f13-948c-8db2b4aad424-kube-api-access-4fjcl\") pod \"40b723c1-2e4c-4f13-948c-8db2b4aad424\" (UID: \"40b723c1-2e4c-4f13-948c-8db2b4aad424\") " Dec 09 10:03:02 crc kubenswrapper[4824]: I1209 10:03:02.626901 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40b723c1-2e4c-4f13-948c-8db2b4aad424-utilities\") pod \"40b723c1-2e4c-4f13-948c-8db2b4aad424\" (UID: \"40b723c1-2e4c-4f13-948c-8db2b4aad424\") " Dec 09 10:03:02 crc kubenswrapper[4824]: I1209 10:03:02.626931 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40b723c1-2e4c-4f13-948c-8db2b4aad424-catalog-content\") pod \"40b723c1-2e4c-4f13-948c-8db2b4aad424\" (UID: \"40b723c1-2e4c-4f13-948c-8db2b4aad424\") " Dec 09 10:03:02 crc kubenswrapper[4824]: I1209 10:03:02.628286 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40b723c1-2e4c-4f13-948c-8db2b4aad424-utilities" (OuterVolumeSpecName: "utilities") pod "40b723c1-2e4c-4f13-948c-8db2b4aad424" (UID: "40b723c1-2e4c-4f13-948c-8db2b4aad424"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:03:02 crc kubenswrapper[4824]: I1209 10:03:02.636460 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40b723c1-2e4c-4f13-948c-8db2b4aad424-kube-api-access-4fjcl" (OuterVolumeSpecName: "kube-api-access-4fjcl") pod "40b723c1-2e4c-4f13-948c-8db2b4aad424" (UID: "40b723c1-2e4c-4f13-948c-8db2b4aad424"). InnerVolumeSpecName "kube-api-access-4fjcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:03:02 crc kubenswrapper[4824]: I1209 10:03:02.648964 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40b723c1-2e4c-4f13-948c-8db2b4aad424-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "40b723c1-2e4c-4f13-948c-8db2b4aad424" (UID: "40b723c1-2e4c-4f13-948c-8db2b4aad424"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:03:02 crc kubenswrapper[4824]: I1209 10:03:02.729523 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fjcl\" (UniqueName: \"kubernetes.io/projected/40b723c1-2e4c-4f13-948c-8db2b4aad424-kube-api-access-4fjcl\") on node \"crc\" DevicePath \"\"" Dec 09 10:03:02 crc kubenswrapper[4824]: I1209 10:03:02.729585 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40b723c1-2e4c-4f13-948c-8db2b4aad424-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 10:03:02 crc kubenswrapper[4824]: I1209 10:03:02.729600 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40b723c1-2e4c-4f13-948c-8db2b4aad424-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 10:03:02 crc kubenswrapper[4824]: I1209 10:03:02.993521 4824 generic.go:334] "Generic (PLEG): container finished" podID="40b723c1-2e4c-4f13-948c-8db2b4aad424" containerID="fd4b3bf21aead16cfe1de302b14bf0e9443790b047c135b46a16adefd149b26b" exitCode=0 Dec 09 10:03:02 crc kubenswrapper[4824]: I1209 10:03:02.993579 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-44svc" event={"ID":"40b723c1-2e4c-4f13-948c-8db2b4aad424","Type":"ContainerDied","Data":"fd4b3bf21aead16cfe1de302b14bf0e9443790b047c135b46a16adefd149b26b"} Dec 09 10:03:02 crc kubenswrapper[4824]: I1209 10:03:02.993618 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-44svc" event={"ID":"40b723c1-2e4c-4f13-948c-8db2b4aad424","Type":"ContainerDied","Data":"3436f0a8ee0b59a457c8e515c565579c2c5dcc71408f58375ae6e7cc42c43ddd"} Dec 09 10:03:02 crc kubenswrapper[4824]: I1209 10:03:02.993643 4824 scope.go:117] "RemoveContainer" containerID="fd4b3bf21aead16cfe1de302b14bf0e9443790b047c135b46a16adefd149b26b" Dec 09 10:03:02 crc kubenswrapper[4824]: I1209 10:03:02.993669 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-44svc" Dec 09 10:03:03 crc kubenswrapper[4824]: I1209 10:03:03.016251 4824 scope.go:117] "RemoveContainer" containerID="f82221ae45609757673eb7d09537d232b45d154559bfc59d1e90e862735dc5e1" Dec 09 10:03:03 crc kubenswrapper[4824]: I1209 10:03:03.035312 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-44svc"] Dec 09 10:03:03 crc kubenswrapper[4824]: I1209 10:03:03.045441 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-44svc"] Dec 09 10:03:03 crc kubenswrapper[4824]: I1209 10:03:03.058824 4824 scope.go:117] "RemoveContainer" containerID="84492123746f740e5ed1f8b21dd84ae821743aa0ea56915950f319a16d70e6e0" Dec 09 10:03:03 crc kubenswrapper[4824]: I1209 10:03:03.078239 4824 scope.go:117] "RemoveContainer" containerID="fd4b3bf21aead16cfe1de302b14bf0e9443790b047c135b46a16adefd149b26b" Dec 09 10:03:03 crc kubenswrapper[4824]: E1209 10:03:03.078741 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd4b3bf21aead16cfe1de302b14bf0e9443790b047c135b46a16adefd149b26b\": container with ID starting with fd4b3bf21aead16cfe1de302b14bf0e9443790b047c135b46a16adefd149b26b not found: ID does not exist" containerID="fd4b3bf21aead16cfe1de302b14bf0e9443790b047c135b46a16adefd149b26b" Dec 09 10:03:03 crc kubenswrapper[4824]: I1209 10:03:03.078820 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd4b3bf21aead16cfe1de302b14bf0e9443790b047c135b46a16adefd149b26b"} err="failed to get container status \"fd4b3bf21aead16cfe1de302b14bf0e9443790b047c135b46a16adefd149b26b\": rpc error: code = NotFound desc = could not find container \"fd4b3bf21aead16cfe1de302b14bf0e9443790b047c135b46a16adefd149b26b\": container with ID starting with fd4b3bf21aead16cfe1de302b14bf0e9443790b047c135b46a16adefd149b26b not found: ID does not exist" Dec 09 10:03:03 crc kubenswrapper[4824]: I1209 10:03:03.078857 4824 scope.go:117] "RemoveContainer" containerID="f82221ae45609757673eb7d09537d232b45d154559bfc59d1e90e862735dc5e1" Dec 09 10:03:03 crc kubenswrapper[4824]: E1209 10:03:03.079242 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f82221ae45609757673eb7d09537d232b45d154559bfc59d1e90e862735dc5e1\": container with ID starting with f82221ae45609757673eb7d09537d232b45d154559bfc59d1e90e862735dc5e1 not found: ID does not exist" containerID="f82221ae45609757673eb7d09537d232b45d154559bfc59d1e90e862735dc5e1" Dec 09 10:03:03 crc kubenswrapper[4824]: I1209 10:03:03.079299 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f82221ae45609757673eb7d09537d232b45d154559bfc59d1e90e862735dc5e1"} err="failed to get container status \"f82221ae45609757673eb7d09537d232b45d154559bfc59d1e90e862735dc5e1\": rpc error: code = NotFound desc = could not find container \"f82221ae45609757673eb7d09537d232b45d154559bfc59d1e90e862735dc5e1\": container with ID starting with f82221ae45609757673eb7d09537d232b45d154559bfc59d1e90e862735dc5e1 not found: ID does not exist" Dec 09 10:03:03 crc kubenswrapper[4824]: I1209 10:03:03.079332 4824 scope.go:117] "RemoveContainer" containerID="84492123746f740e5ed1f8b21dd84ae821743aa0ea56915950f319a16d70e6e0" Dec 09 10:03:03 crc kubenswrapper[4824]: E1209 10:03:03.079604 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84492123746f740e5ed1f8b21dd84ae821743aa0ea56915950f319a16d70e6e0\": container with ID starting with 84492123746f740e5ed1f8b21dd84ae821743aa0ea56915950f319a16d70e6e0 not found: ID does not exist" containerID="84492123746f740e5ed1f8b21dd84ae821743aa0ea56915950f319a16d70e6e0" Dec 09 10:03:03 crc kubenswrapper[4824]: I1209 10:03:03.079637 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84492123746f740e5ed1f8b21dd84ae821743aa0ea56915950f319a16d70e6e0"} err="failed to get container status \"84492123746f740e5ed1f8b21dd84ae821743aa0ea56915950f319a16d70e6e0\": rpc error: code = NotFound desc = could not find container \"84492123746f740e5ed1f8b21dd84ae821743aa0ea56915950f319a16d70e6e0\": container with ID starting with 84492123746f740e5ed1f8b21dd84ae821743aa0ea56915950f319a16d70e6e0 not found: ID does not exist" Dec 09 10:03:03 crc kubenswrapper[4824]: I1209 10:03:03.922669 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40b723c1-2e4c-4f13-948c-8db2b4aad424" path="/var/lib/kubelet/pods/40b723c1-2e4c-4f13-948c-8db2b4aad424/volumes" Dec 09 10:03:09 crc kubenswrapper[4824]: I1209 10:03:09.959110 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wjlqj"] Dec 09 10:03:09 crc kubenswrapper[4824]: E1209 10:03:09.960180 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40b723c1-2e4c-4f13-948c-8db2b4aad424" containerName="registry-server" Dec 09 10:03:09 crc kubenswrapper[4824]: I1209 10:03:09.960202 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="40b723c1-2e4c-4f13-948c-8db2b4aad424" containerName="registry-server" Dec 09 10:03:09 crc kubenswrapper[4824]: E1209 10:03:09.960217 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290" containerName="registry-server" Dec 09 10:03:09 crc kubenswrapper[4824]: I1209 10:03:09.960224 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290" containerName="registry-server" Dec 09 10:03:09 crc kubenswrapper[4824]: E1209 10:03:09.960259 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40b723c1-2e4c-4f13-948c-8db2b4aad424" containerName="extract-content" Dec 09 10:03:09 crc kubenswrapper[4824]: I1209 10:03:09.960268 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="40b723c1-2e4c-4f13-948c-8db2b4aad424" containerName="extract-content" Dec 09 10:03:09 crc kubenswrapper[4824]: E1209 10:03:09.960291 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290" containerName="extract-content" Dec 09 10:03:09 crc kubenswrapper[4824]: I1209 10:03:09.960765 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290" containerName="extract-content" Dec 09 10:03:09 crc kubenswrapper[4824]: E1209 10:03:09.960796 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40b723c1-2e4c-4f13-948c-8db2b4aad424" containerName="extract-utilities" Dec 09 10:03:09 crc kubenswrapper[4824]: I1209 10:03:09.960805 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="40b723c1-2e4c-4f13-948c-8db2b4aad424" containerName="extract-utilities" Dec 09 10:03:09 crc kubenswrapper[4824]: E1209 10:03:09.960820 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290" containerName="extract-utilities" Dec 09 10:03:09 crc kubenswrapper[4824]: I1209 10:03:09.960827 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290" containerName="extract-utilities" Dec 09 10:03:09 crc kubenswrapper[4824]: I1209 10:03:09.961011 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="40b723c1-2e4c-4f13-948c-8db2b4aad424" containerName="registry-server" Dec 09 10:03:09 crc kubenswrapper[4824]: I1209 10:03:09.961029 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d49df1d-a9d9-4e2a-bfec-4b0dc4be0290" containerName="registry-server" Dec 09 10:03:09 crc kubenswrapper[4824]: I1209 10:03:09.962349 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wjlqj" Dec 09 10:03:09 crc kubenswrapper[4824]: I1209 10:03:09.972675 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wjlqj"] Dec 09 10:03:10 crc kubenswrapper[4824]: I1209 10:03:10.066376 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4932684f-7dba-4984-9fa6-3abb519e814d-catalog-content\") pod \"certified-operators-wjlqj\" (UID: \"4932684f-7dba-4984-9fa6-3abb519e814d\") " pod="openshift-marketplace/certified-operators-wjlqj" Dec 09 10:03:10 crc kubenswrapper[4824]: I1209 10:03:10.066447 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4932684f-7dba-4984-9fa6-3abb519e814d-utilities\") pod \"certified-operators-wjlqj\" (UID: \"4932684f-7dba-4984-9fa6-3abb519e814d\") " pod="openshift-marketplace/certified-operators-wjlqj" Dec 09 10:03:10 crc kubenswrapper[4824]: I1209 10:03:10.066477 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shfhx\" (UniqueName: \"kubernetes.io/projected/4932684f-7dba-4984-9fa6-3abb519e814d-kube-api-access-shfhx\") pod \"certified-operators-wjlqj\" (UID: \"4932684f-7dba-4984-9fa6-3abb519e814d\") " pod="openshift-marketplace/certified-operators-wjlqj" Dec 09 10:03:10 crc kubenswrapper[4824]: I1209 10:03:10.168002 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4932684f-7dba-4984-9fa6-3abb519e814d-catalog-content\") pod \"certified-operators-wjlqj\" (UID: \"4932684f-7dba-4984-9fa6-3abb519e814d\") " pod="openshift-marketplace/certified-operators-wjlqj" Dec 09 10:03:10 crc kubenswrapper[4824]: I1209 10:03:10.168072 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4932684f-7dba-4984-9fa6-3abb519e814d-utilities\") pod \"certified-operators-wjlqj\" (UID: \"4932684f-7dba-4984-9fa6-3abb519e814d\") " pod="openshift-marketplace/certified-operators-wjlqj" Dec 09 10:03:10 crc kubenswrapper[4824]: I1209 10:03:10.168099 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shfhx\" (UniqueName: \"kubernetes.io/projected/4932684f-7dba-4984-9fa6-3abb519e814d-kube-api-access-shfhx\") pod \"certified-operators-wjlqj\" (UID: \"4932684f-7dba-4984-9fa6-3abb519e814d\") " pod="openshift-marketplace/certified-operators-wjlqj" Dec 09 10:03:10 crc kubenswrapper[4824]: I1209 10:03:10.168834 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4932684f-7dba-4984-9fa6-3abb519e814d-catalog-content\") pod \"certified-operators-wjlqj\" (UID: \"4932684f-7dba-4984-9fa6-3abb519e814d\") " pod="openshift-marketplace/certified-operators-wjlqj" Dec 09 10:03:10 crc kubenswrapper[4824]: I1209 10:03:10.168857 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4932684f-7dba-4984-9fa6-3abb519e814d-utilities\") pod \"certified-operators-wjlqj\" (UID: \"4932684f-7dba-4984-9fa6-3abb519e814d\") " pod="openshift-marketplace/certified-operators-wjlqj" Dec 09 10:03:10 crc kubenswrapper[4824]: I1209 10:03:10.202666 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shfhx\" (UniqueName: \"kubernetes.io/projected/4932684f-7dba-4984-9fa6-3abb519e814d-kube-api-access-shfhx\") pod \"certified-operators-wjlqj\" (UID: \"4932684f-7dba-4984-9fa6-3abb519e814d\") " pod="openshift-marketplace/certified-operators-wjlqj" Dec 09 10:03:10 crc kubenswrapper[4824]: I1209 10:03:10.292015 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wjlqj" Dec 09 10:03:10 crc kubenswrapper[4824]: I1209 10:03:10.881730 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wjlqj"] Dec 09 10:03:11 crc kubenswrapper[4824]: I1209 10:03:11.062550 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wjlqj" event={"ID":"4932684f-7dba-4984-9fa6-3abb519e814d","Type":"ContainerStarted","Data":"f881d07e6b7f3afe2411f6aba644ee24f2fbb2c29305e79b2848d62cbc4eba1c"} Dec 09 10:03:12 crc kubenswrapper[4824]: I1209 10:03:12.085947 4824 generic.go:334] "Generic (PLEG): container finished" podID="4932684f-7dba-4984-9fa6-3abb519e814d" containerID="7a4171a86461ddd4139fa3352f0f92b253b15b25e80b8b5d95dd0a328602ef94" exitCode=0 Dec 09 10:03:12 crc kubenswrapper[4824]: I1209 10:03:12.086607 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wjlqj" event={"ID":"4932684f-7dba-4984-9fa6-3abb519e814d","Type":"ContainerDied","Data":"7a4171a86461ddd4139fa3352f0f92b253b15b25e80b8b5d95dd0a328602ef94"} Dec 09 10:03:13 crc kubenswrapper[4824]: I1209 10:03:13.098173 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wjlqj" event={"ID":"4932684f-7dba-4984-9fa6-3abb519e814d","Type":"ContainerStarted","Data":"401cde0d07c9ce1ea4ba163bc0eed0d9c419ae1f4fa3228037fdddaaf68c50ae"} Dec 09 10:03:14 crc kubenswrapper[4824]: I1209 10:03:14.108999 4824 generic.go:334] "Generic (PLEG): container finished" podID="4932684f-7dba-4984-9fa6-3abb519e814d" containerID="401cde0d07c9ce1ea4ba163bc0eed0d9c419ae1f4fa3228037fdddaaf68c50ae" exitCode=0 Dec 09 10:03:14 crc kubenswrapper[4824]: I1209 10:03:14.109590 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wjlqj" event={"ID":"4932684f-7dba-4984-9fa6-3abb519e814d","Type":"ContainerDied","Data":"401cde0d07c9ce1ea4ba163bc0eed0d9c419ae1f4fa3228037fdddaaf68c50ae"} Dec 09 10:03:16 crc kubenswrapper[4824]: I1209 10:03:16.137351 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wjlqj" event={"ID":"4932684f-7dba-4984-9fa6-3abb519e814d","Type":"ContainerStarted","Data":"e3897ceaf39eca4895e7451747f84efbaa9500445b545319b8130d6fed3b79ec"} Dec 09 10:03:16 crc kubenswrapper[4824]: I1209 10:03:16.177829 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wjlqj" podStartSLOduration=4.185769195 podStartE2EDuration="7.177802975s" podCreationTimestamp="2025-12-09 10:03:09 +0000 UTC" firstStartedPulling="2025-12-09 10:03:12.090364971 +0000 UTC m=+948.424869638" lastFinishedPulling="2025-12-09 10:03:15.082398751 +0000 UTC m=+951.416903418" observedRunningTime="2025-12-09 10:03:16.166126336 +0000 UTC m=+952.500631013" watchObservedRunningTime="2025-12-09 10:03:16.177802975 +0000 UTC m=+952.512307642" Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:16.999888 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-79d58c8b9c-dbzmt" podUID="065d385c-89bf-409f-822e-58652c09a9a3" containerName="console" containerID="cri-o://d675c1d6889d781b51f200d99f17df80a14f3b02b3443ab8758e818dd96307f6" gracePeriod=15 Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.148201 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-79d58c8b9c-dbzmt_065d385c-89bf-409f-822e-58652c09a9a3/console/0.log" Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.148532 4824 generic.go:334] "Generic (PLEG): container finished" podID="065d385c-89bf-409f-822e-58652c09a9a3" containerID="d675c1d6889d781b51f200d99f17df80a14f3b02b3443ab8758e818dd96307f6" exitCode=2 Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.148635 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-79d58c8b9c-dbzmt" event={"ID":"065d385c-89bf-409f-822e-58652c09a9a3","Type":"ContainerDied","Data":"d675c1d6889d781b51f200d99f17df80a14f3b02b3443ab8758e818dd96307f6"} Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.486256 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-79d58c8b9c-dbzmt_065d385c-89bf-409f-822e-58652c09a9a3/console/0.log" Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.486629 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.522318 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/065d385c-89bf-409f-822e-58652c09a9a3-trusted-ca-bundle\") pod \"065d385c-89bf-409f-822e-58652c09a9a3\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.522462 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/065d385c-89bf-409f-822e-58652c09a9a3-console-serving-cert\") pod \"065d385c-89bf-409f-822e-58652c09a9a3\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.522523 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fl5v8\" (UniqueName: \"kubernetes.io/projected/065d385c-89bf-409f-822e-58652c09a9a3-kube-api-access-fl5v8\") pod \"065d385c-89bf-409f-822e-58652c09a9a3\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.522595 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/065d385c-89bf-409f-822e-58652c09a9a3-service-ca\") pod \"065d385c-89bf-409f-822e-58652c09a9a3\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.522628 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/065d385c-89bf-409f-822e-58652c09a9a3-console-config\") pod \"065d385c-89bf-409f-822e-58652c09a9a3\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.522671 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/065d385c-89bf-409f-822e-58652c09a9a3-oauth-serving-cert\") pod \"065d385c-89bf-409f-822e-58652c09a9a3\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.522692 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/065d385c-89bf-409f-822e-58652c09a9a3-console-oauth-config\") pod \"065d385c-89bf-409f-822e-58652c09a9a3\" (UID: \"065d385c-89bf-409f-822e-58652c09a9a3\") " Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.523738 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/065d385c-89bf-409f-822e-58652c09a9a3-console-config" (OuterVolumeSpecName: "console-config") pod "065d385c-89bf-409f-822e-58652c09a9a3" (UID: "065d385c-89bf-409f-822e-58652c09a9a3"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.523981 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/065d385c-89bf-409f-822e-58652c09a9a3-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "065d385c-89bf-409f-822e-58652c09a9a3" (UID: "065d385c-89bf-409f-822e-58652c09a9a3"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.524224 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/065d385c-89bf-409f-822e-58652c09a9a3-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "065d385c-89bf-409f-822e-58652c09a9a3" (UID: "065d385c-89bf-409f-822e-58652c09a9a3"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.525333 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/065d385c-89bf-409f-822e-58652c09a9a3-service-ca" (OuterVolumeSpecName: "service-ca") pod "065d385c-89bf-409f-822e-58652c09a9a3" (UID: "065d385c-89bf-409f-822e-58652c09a9a3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.543396 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/065d385c-89bf-409f-822e-58652c09a9a3-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "065d385c-89bf-409f-822e-58652c09a9a3" (UID: "065d385c-89bf-409f-822e-58652c09a9a3"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.543540 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/065d385c-89bf-409f-822e-58652c09a9a3-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "065d385c-89bf-409f-822e-58652c09a9a3" (UID: "065d385c-89bf-409f-822e-58652c09a9a3"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.543929 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/065d385c-89bf-409f-822e-58652c09a9a3-kube-api-access-fl5v8" (OuterVolumeSpecName: "kube-api-access-fl5v8") pod "065d385c-89bf-409f-822e-58652c09a9a3" (UID: "065d385c-89bf-409f-822e-58652c09a9a3"). InnerVolumeSpecName "kube-api-access-fl5v8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.625129 4824 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/065d385c-89bf-409f-822e-58652c09a9a3-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.625184 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fl5v8\" (UniqueName: \"kubernetes.io/projected/065d385c-89bf-409f-822e-58652c09a9a3-kube-api-access-fl5v8\") on node \"crc\" DevicePath \"\"" Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.625203 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/065d385c-89bf-409f-822e-58652c09a9a3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.625215 4824 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/065d385c-89bf-409f-822e-58652c09a9a3-console-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.625227 4824 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/065d385c-89bf-409f-822e-58652c09a9a3-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.625239 4824 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/065d385c-89bf-409f-822e-58652c09a9a3-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:03:17 crc kubenswrapper[4824]: I1209 10:03:17.625249 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/065d385c-89bf-409f-822e-58652c09a9a3-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:03:18 crc kubenswrapper[4824]: I1209 10:03:18.162146 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-79d58c8b9c-dbzmt_065d385c-89bf-409f-822e-58652c09a9a3/console/0.log" Dec 09 10:03:18 crc kubenswrapper[4824]: I1209 10:03:18.163410 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-79d58c8b9c-dbzmt" event={"ID":"065d385c-89bf-409f-822e-58652c09a9a3","Type":"ContainerDied","Data":"7e0db67401027fc28ff91cf556926cd41abb94b9bff90553a0f5f023c6631532"} Dec 09 10:03:18 crc kubenswrapper[4824]: I1209 10:03:18.163456 4824 scope.go:117] "RemoveContainer" containerID="d675c1d6889d781b51f200d99f17df80a14f3b02b3443ab8758e818dd96307f6" Dec 09 10:03:18 crc kubenswrapper[4824]: I1209 10:03:18.163497 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-79d58c8b9c-dbzmt" Dec 09 10:03:18 crc kubenswrapper[4824]: I1209 10:03:18.192096 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-79d58c8b9c-dbzmt"] Dec 09 10:03:18 crc kubenswrapper[4824]: I1209 10:03:18.199721 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-79d58c8b9c-dbzmt"] Dec 09 10:03:19 crc kubenswrapper[4824]: I1209 10:03:19.920465 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="065d385c-89bf-409f-822e-58652c09a9a3" path="/var/lib/kubelet/pods/065d385c-89bf-409f-822e-58652c09a9a3/volumes" Dec 09 10:03:20 crc kubenswrapper[4824]: I1209 10:03:20.292438 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wjlqj" Dec 09 10:03:20 crc kubenswrapper[4824]: I1209 10:03:20.292932 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wjlqj" Dec 09 10:03:20 crc kubenswrapper[4824]: I1209 10:03:20.351250 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wjlqj" Dec 09 10:03:21 crc kubenswrapper[4824]: I1209 10:03:21.264211 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wjlqj" Dec 09 10:03:21 crc kubenswrapper[4824]: I1209 10:03:21.384819 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t"] Dec 09 10:03:21 crc kubenswrapper[4824]: E1209 10:03:21.385238 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="065d385c-89bf-409f-822e-58652c09a9a3" containerName="console" Dec 09 10:03:21 crc kubenswrapper[4824]: I1209 10:03:21.385262 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="065d385c-89bf-409f-822e-58652c09a9a3" containerName="console" Dec 09 10:03:21 crc kubenswrapper[4824]: I1209 10:03:21.385438 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="065d385c-89bf-409f-822e-58652c09a9a3" containerName="console" Dec 09 10:03:21 crc kubenswrapper[4824]: I1209 10:03:21.386842 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t" Dec 09 10:03:21 crc kubenswrapper[4824]: I1209 10:03:21.390718 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 09 10:03:21 crc kubenswrapper[4824]: I1209 10:03:21.409553 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t"] Dec 09 10:03:21 crc kubenswrapper[4824]: I1209 10:03:21.494728 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44aedeb1-3d60-43f0-9a8b-472e489077e0-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t\" (UID: \"44aedeb1-3d60-43f0-9a8b-472e489077e0\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t" Dec 09 10:03:21 crc kubenswrapper[4824]: I1209 10:03:21.494925 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqrhp\" (UniqueName: \"kubernetes.io/projected/44aedeb1-3d60-43f0-9a8b-472e489077e0-kube-api-access-fqrhp\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t\" (UID: \"44aedeb1-3d60-43f0-9a8b-472e489077e0\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t" Dec 09 10:03:21 crc kubenswrapper[4824]: I1209 10:03:21.494986 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44aedeb1-3d60-43f0-9a8b-472e489077e0-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t\" (UID: \"44aedeb1-3d60-43f0-9a8b-472e489077e0\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t" Dec 09 10:03:21 crc kubenswrapper[4824]: I1209 10:03:21.597072 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqrhp\" (UniqueName: \"kubernetes.io/projected/44aedeb1-3d60-43f0-9a8b-472e489077e0-kube-api-access-fqrhp\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t\" (UID: \"44aedeb1-3d60-43f0-9a8b-472e489077e0\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t" Dec 09 10:03:21 crc kubenswrapper[4824]: I1209 10:03:21.597167 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44aedeb1-3d60-43f0-9a8b-472e489077e0-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t\" (UID: \"44aedeb1-3d60-43f0-9a8b-472e489077e0\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t" Dec 09 10:03:21 crc kubenswrapper[4824]: I1209 10:03:21.597218 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44aedeb1-3d60-43f0-9a8b-472e489077e0-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t\" (UID: \"44aedeb1-3d60-43f0-9a8b-472e489077e0\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t" Dec 09 10:03:21 crc kubenswrapper[4824]: I1209 10:03:21.597803 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44aedeb1-3d60-43f0-9a8b-472e489077e0-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t\" (UID: \"44aedeb1-3d60-43f0-9a8b-472e489077e0\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t" Dec 09 10:03:21 crc kubenswrapper[4824]: I1209 10:03:21.597844 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44aedeb1-3d60-43f0-9a8b-472e489077e0-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t\" (UID: \"44aedeb1-3d60-43f0-9a8b-472e489077e0\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t" Dec 09 10:03:21 crc kubenswrapper[4824]: I1209 10:03:21.628515 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqrhp\" (UniqueName: \"kubernetes.io/projected/44aedeb1-3d60-43f0-9a8b-472e489077e0-kube-api-access-fqrhp\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t\" (UID: \"44aedeb1-3d60-43f0-9a8b-472e489077e0\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t" Dec 09 10:03:21 crc kubenswrapper[4824]: I1209 10:03:21.714578 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t" Dec 09 10:03:22 crc kubenswrapper[4824]: I1209 10:03:22.005397 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t"] Dec 09 10:03:22 crc kubenswrapper[4824]: I1209 10:03:22.201298 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t" event={"ID":"44aedeb1-3d60-43f0-9a8b-472e489077e0","Type":"ContainerStarted","Data":"d100ca689b005aebcbdf5144e6e47231e5c346431c7e4dee6e96f2d9f95716d6"} Dec 09 10:03:23 crc kubenswrapper[4824]: I1209 10:03:23.211563 4824 generic.go:334] "Generic (PLEG): container finished" podID="44aedeb1-3d60-43f0-9a8b-472e489077e0" containerID="ce4be388d2af4540bf807b4c77e628c63e0e600737c1695fb56bc276b49608a2" exitCode=0 Dec 09 10:03:23 crc kubenswrapper[4824]: I1209 10:03:23.211630 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t" event={"ID":"44aedeb1-3d60-43f0-9a8b-472e489077e0","Type":"ContainerDied","Data":"ce4be388d2af4540bf807b4c77e628c63e0e600737c1695fb56bc276b49608a2"} Dec 09 10:03:24 crc kubenswrapper[4824]: I1209 10:03:24.589038 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wjlqj"] Dec 09 10:03:24 crc kubenswrapper[4824]: I1209 10:03:24.589728 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wjlqj" podUID="4932684f-7dba-4984-9fa6-3abb519e814d" containerName="registry-server" containerID="cri-o://e3897ceaf39eca4895e7451747f84efbaa9500445b545319b8130d6fed3b79ec" gracePeriod=2 Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.075529 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wjlqj" Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.211070 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shfhx\" (UniqueName: \"kubernetes.io/projected/4932684f-7dba-4984-9fa6-3abb519e814d-kube-api-access-shfhx\") pod \"4932684f-7dba-4984-9fa6-3abb519e814d\" (UID: \"4932684f-7dba-4984-9fa6-3abb519e814d\") " Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.211184 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4932684f-7dba-4984-9fa6-3abb519e814d-catalog-content\") pod \"4932684f-7dba-4984-9fa6-3abb519e814d\" (UID: \"4932684f-7dba-4984-9fa6-3abb519e814d\") " Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.211218 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4932684f-7dba-4984-9fa6-3abb519e814d-utilities\") pod \"4932684f-7dba-4984-9fa6-3abb519e814d\" (UID: \"4932684f-7dba-4984-9fa6-3abb519e814d\") " Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.212538 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4932684f-7dba-4984-9fa6-3abb519e814d-utilities" (OuterVolumeSpecName: "utilities") pod "4932684f-7dba-4984-9fa6-3abb519e814d" (UID: "4932684f-7dba-4984-9fa6-3abb519e814d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.220306 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4932684f-7dba-4984-9fa6-3abb519e814d-kube-api-access-shfhx" (OuterVolumeSpecName: "kube-api-access-shfhx") pod "4932684f-7dba-4984-9fa6-3abb519e814d" (UID: "4932684f-7dba-4984-9fa6-3abb519e814d"). InnerVolumeSpecName "kube-api-access-shfhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.232295 4824 generic.go:334] "Generic (PLEG): container finished" podID="4932684f-7dba-4984-9fa6-3abb519e814d" containerID="e3897ceaf39eca4895e7451747f84efbaa9500445b545319b8130d6fed3b79ec" exitCode=0 Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.232361 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wjlqj" Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.232393 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wjlqj" event={"ID":"4932684f-7dba-4984-9fa6-3abb519e814d","Type":"ContainerDied","Data":"e3897ceaf39eca4895e7451747f84efbaa9500445b545319b8130d6fed3b79ec"} Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.232456 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wjlqj" event={"ID":"4932684f-7dba-4984-9fa6-3abb519e814d","Type":"ContainerDied","Data":"f881d07e6b7f3afe2411f6aba644ee24f2fbb2c29305e79b2848d62cbc4eba1c"} Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.232516 4824 scope.go:117] "RemoveContainer" containerID="e3897ceaf39eca4895e7451747f84efbaa9500445b545319b8130d6fed3b79ec" Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.234840 4824 generic.go:334] "Generic (PLEG): container finished" podID="44aedeb1-3d60-43f0-9a8b-472e489077e0" containerID="317ddb9f3c2e6c2877cc4d9c14db8b06efd47b9b084294d7d6695ee424ebf798" exitCode=0 Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.234873 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t" event={"ID":"44aedeb1-3d60-43f0-9a8b-472e489077e0","Type":"ContainerDied","Data":"317ddb9f3c2e6c2877cc4d9c14db8b06efd47b9b084294d7d6695ee424ebf798"} Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.284162 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4932684f-7dba-4984-9fa6-3abb519e814d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4932684f-7dba-4984-9fa6-3abb519e814d" (UID: "4932684f-7dba-4984-9fa6-3abb519e814d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.285321 4824 scope.go:117] "RemoveContainer" containerID="401cde0d07c9ce1ea4ba163bc0eed0d9c419ae1f4fa3228037fdddaaf68c50ae" Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.313650 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shfhx\" (UniqueName: \"kubernetes.io/projected/4932684f-7dba-4984-9fa6-3abb519e814d-kube-api-access-shfhx\") on node \"crc\" DevicePath \"\"" Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.313693 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4932684f-7dba-4984-9fa6-3abb519e814d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.313702 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4932684f-7dba-4984-9fa6-3abb519e814d-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.327319 4824 scope.go:117] "RemoveContainer" containerID="7a4171a86461ddd4139fa3352f0f92b253b15b25e80b8b5d95dd0a328602ef94" Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.357894 4824 scope.go:117] "RemoveContainer" containerID="e3897ceaf39eca4895e7451747f84efbaa9500445b545319b8130d6fed3b79ec" Dec 09 10:03:25 crc kubenswrapper[4824]: E1209 10:03:25.358458 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3897ceaf39eca4895e7451747f84efbaa9500445b545319b8130d6fed3b79ec\": container with ID starting with e3897ceaf39eca4895e7451747f84efbaa9500445b545319b8130d6fed3b79ec not found: ID does not exist" containerID="e3897ceaf39eca4895e7451747f84efbaa9500445b545319b8130d6fed3b79ec" Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.358531 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3897ceaf39eca4895e7451747f84efbaa9500445b545319b8130d6fed3b79ec"} err="failed to get container status \"e3897ceaf39eca4895e7451747f84efbaa9500445b545319b8130d6fed3b79ec\": rpc error: code = NotFound desc = could not find container \"e3897ceaf39eca4895e7451747f84efbaa9500445b545319b8130d6fed3b79ec\": container with ID starting with e3897ceaf39eca4895e7451747f84efbaa9500445b545319b8130d6fed3b79ec not found: ID does not exist" Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.358570 4824 scope.go:117] "RemoveContainer" containerID="401cde0d07c9ce1ea4ba163bc0eed0d9c419ae1f4fa3228037fdddaaf68c50ae" Dec 09 10:03:25 crc kubenswrapper[4824]: E1209 10:03:25.359008 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"401cde0d07c9ce1ea4ba163bc0eed0d9c419ae1f4fa3228037fdddaaf68c50ae\": container with ID starting with 401cde0d07c9ce1ea4ba163bc0eed0d9c419ae1f4fa3228037fdddaaf68c50ae not found: ID does not exist" containerID="401cde0d07c9ce1ea4ba163bc0eed0d9c419ae1f4fa3228037fdddaaf68c50ae" Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.359065 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"401cde0d07c9ce1ea4ba163bc0eed0d9c419ae1f4fa3228037fdddaaf68c50ae"} err="failed to get container status \"401cde0d07c9ce1ea4ba163bc0eed0d9c419ae1f4fa3228037fdddaaf68c50ae\": rpc error: code = NotFound desc = could not find container \"401cde0d07c9ce1ea4ba163bc0eed0d9c419ae1f4fa3228037fdddaaf68c50ae\": container with ID starting with 401cde0d07c9ce1ea4ba163bc0eed0d9c419ae1f4fa3228037fdddaaf68c50ae not found: ID does not exist" Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.359101 4824 scope.go:117] "RemoveContainer" containerID="7a4171a86461ddd4139fa3352f0f92b253b15b25e80b8b5d95dd0a328602ef94" Dec 09 10:03:25 crc kubenswrapper[4824]: E1209 10:03:25.360432 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a4171a86461ddd4139fa3352f0f92b253b15b25e80b8b5d95dd0a328602ef94\": container with ID starting with 7a4171a86461ddd4139fa3352f0f92b253b15b25e80b8b5d95dd0a328602ef94 not found: ID does not exist" containerID="7a4171a86461ddd4139fa3352f0f92b253b15b25e80b8b5d95dd0a328602ef94" Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.360498 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a4171a86461ddd4139fa3352f0f92b253b15b25e80b8b5d95dd0a328602ef94"} err="failed to get container status \"7a4171a86461ddd4139fa3352f0f92b253b15b25e80b8b5d95dd0a328602ef94\": rpc error: code = NotFound desc = could not find container \"7a4171a86461ddd4139fa3352f0f92b253b15b25e80b8b5d95dd0a328602ef94\": container with ID starting with 7a4171a86461ddd4139fa3352f0f92b253b15b25e80b8b5d95dd0a328602ef94 not found: ID does not exist" Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.612088 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wjlqj"] Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.624665 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wjlqj"] Dec 09 10:03:25 crc kubenswrapper[4824]: I1209 10:03:25.925892 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4932684f-7dba-4984-9fa6-3abb519e814d" path="/var/lib/kubelet/pods/4932684f-7dba-4984-9fa6-3abb519e814d/volumes" Dec 09 10:03:26 crc kubenswrapper[4824]: I1209 10:03:26.246736 4824 generic.go:334] "Generic (PLEG): container finished" podID="44aedeb1-3d60-43f0-9a8b-472e489077e0" containerID="95482d561e672899c2f61a2bac7f18a1e9d901c486aabd14956a2744c6b55810" exitCode=0 Dec 09 10:03:26 crc kubenswrapper[4824]: I1209 10:03:26.246812 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t" event={"ID":"44aedeb1-3d60-43f0-9a8b-472e489077e0","Type":"ContainerDied","Data":"95482d561e672899c2f61a2bac7f18a1e9d901c486aabd14956a2744c6b55810"} Dec 09 10:03:27 crc kubenswrapper[4824]: I1209 10:03:27.612234 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t" Dec 09 10:03:27 crc kubenswrapper[4824]: I1209 10:03:27.761356 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqrhp\" (UniqueName: \"kubernetes.io/projected/44aedeb1-3d60-43f0-9a8b-472e489077e0-kube-api-access-fqrhp\") pod \"44aedeb1-3d60-43f0-9a8b-472e489077e0\" (UID: \"44aedeb1-3d60-43f0-9a8b-472e489077e0\") " Dec 09 10:03:27 crc kubenswrapper[4824]: I1209 10:03:27.761581 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44aedeb1-3d60-43f0-9a8b-472e489077e0-bundle\") pod \"44aedeb1-3d60-43f0-9a8b-472e489077e0\" (UID: \"44aedeb1-3d60-43f0-9a8b-472e489077e0\") " Dec 09 10:03:27 crc kubenswrapper[4824]: I1209 10:03:27.761644 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44aedeb1-3d60-43f0-9a8b-472e489077e0-util\") pod \"44aedeb1-3d60-43f0-9a8b-472e489077e0\" (UID: \"44aedeb1-3d60-43f0-9a8b-472e489077e0\") " Dec 09 10:03:27 crc kubenswrapper[4824]: I1209 10:03:27.762977 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44aedeb1-3d60-43f0-9a8b-472e489077e0-bundle" (OuterVolumeSpecName: "bundle") pod "44aedeb1-3d60-43f0-9a8b-472e489077e0" (UID: "44aedeb1-3d60-43f0-9a8b-472e489077e0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:03:27 crc kubenswrapper[4824]: I1209 10:03:27.776888 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44aedeb1-3d60-43f0-9a8b-472e489077e0-util" (OuterVolumeSpecName: "util") pod "44aedeb1-3d60-43f0-9a8b-472e489077e0" (UID: "44aedeb1-3d60-43f0-9a8b-472e489077e0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:03:27 crc kubenswrapper[4824]: I1209 10:03:27.780718 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44aedeb1-3d60-43f0-9a8b-472e489077e0-kube-api-access-fqrhp" (OuterVolumeSpecName: "kube-api-access-fqrhp") pod "44aedeb1-3d60-43f0-9a8b-472e489077e0" (UID: "44aedeb1-3d60-43f0-9a8b-472e489077e0"). InnerVolumeSpecName "kube-api-access-fqrhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:03:27 crc kubenswrapper[4824]: I1209 10:03:27.863434 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqrhp\" (UniqueName: \"kubernetes.io/projected/44aedeb1-3d60-43f0-9a8b-472e489077e0-kube-api-access-fqrhp\") on node \"crc\" DevicePath \"\"" Dec 09 10:03:27 crc kubenswrapper[4824]: I1209 10:03:27.863476 4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44aedeb1-3d60-43f0-9a8b-472e489077e0-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:03:27 crc kubenswrapper[4824]: I1209 10:03:27.863486 4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44aedeb1-3d60-43f0-9a8b-472e489077e0-util\") on node \"crc\" DevicePath \"\"" Dec 09 10:03:28 crc kubenswrapper[4824]: I1209 10:03:28.263732 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t" event={"ID":"44aedeb1-3d60-43f0-9a8b-472e489077e0","Type":"ContainerDied","Data":"d100ca689b005aebcbdf5144e6e47231e5c346431c7e4dee6e96f2d9f95716d6"} Dec 09 10:03:28 crc kubenswrapper[4824]: I1209 10:03:28.263801 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d100ca689b005aebcbdf5144e6e47231e5c346431c7e4dee6e96f2d9f95716d6" Dec 09 10:03:28 crc kubenswrapper[4824]: I1209 10:03:28.263813 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.066812 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-69bb78d57c-p4l9h"] Dec 09 10:03:39 crc kubenswrapper[4824]: E1209 10:03:39.067961 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4932684f-7dba-4984-9fa6-3abb519e814d" containerName="extract-content" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.067980 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4932684f-7dba-4984-9fa6-3abb519e814d" containerName="extract-content" Dec 09 10:03:39 crc kubenswrapper[4824]: E1209 10:03:39.067994 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44aedeb1-3d60-43f0-9a8b-472e489077e0" containerName="util" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.068002 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="44aedeb1-3d60-43f0-9a8b-472e489077e0" containerName="util" Dec 09 10:03:39 crc kubenswrapper[4824]: E1209 10:03:39.068026 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4932684f-7dba-4984-9fa6-3abb519e814d" containerName="registry-server" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.068036 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4932684f-7dba-4984-9fa6-3abb519e814d" containerName="registry-server" Dec 09 10:03:39 crc kubenswrapper[4824]: E1209 10:03:39.068046 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44aedeb1-3d60-43f0-9a8b-472e489077e0" containerName="extract" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.068064 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="44aedeb1-3d60-43f0-9a8b-472e489077e0" containerName="extract" Dec 09 10:03:39 crc kubenswrapper[4824]: E1209 10:03:39.068080 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4932684f-7dba-4984-9fa6-3abb519e814d" containerName="extract-utilities" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.068087 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4932684f-7dba-4984-9fa6-3abb519e814d" containerName="extract-utilities" Dec 09 10:03:39 crc kubenswrapper[4824]: E1209 10:03:39.068099 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44aedeb1-3d60-43f0-9a8b-472e489077e0" containerName="pull" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.068106 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="44aedeb1-3d60-43f0-9a8b-472e489077e0" containerName="pull" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.068308 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="44aedeb1-3d60-43f0-9a8b-472e489077e0" containerName="extract" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.068322 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4932684f-7dba-4984-9fa6-3abb519e814d" containerName="registry-server" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.069010 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-69bb78d57c-p4l9h" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.071687 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.071695 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.071970 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-g5tsx" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.072217 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.074633 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.088163 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-69bb78d57c-p4l9h"] Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.167072 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz8wf\" (UniqueName: \"kubernetes.io/projected/2fed244b-a135-4d95-a5e5-2eea4bcde7cb-kube-api-access-tz8wf\") pod \"metallb-operator-controller-manager-69bb78d57c-p4l9h\" (UID: \"2fed244b-a135-4d95-a5e5-2eea4bcde7cb\") " pod="metallb-system/metallb-operator-controller-manager-69bb78d57c-p4l9h" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.167291 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2fed244b-a135-4d95-a5e5-2eea4bcde7cb-apiservice-cert\") pod \"metallb-operator-controller-manager-69bb78d57c-p4l9h\" (UID: \"2fed244b-a135-4d95-a5e5-2eea4bcde7cb\") " pod="metallb-system/metallb-operator-controller-manager-69bb78d57c-p4l9h" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.167491 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2fed244b-a135-4d95-a5e5-2eea4bcde7cb-webhook-cert\") pod \"metallb-operator-controller-manager-69bb78d57c-p4l9h\" (UID: \"2fed244b-a135-4d95-a5e5-2eea4bcde7cb\") " pod="metallb-system/metallb-operator-controller-manager-69bb78d57c-p4l9h" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.269895 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz8wf\" (UniqueName: \"kubernetes.io/projected/2fed244b-a135-4d95-a5e5-2eea4bcde7cb-kube-api-access-tz8wf\") pod \"metallb-operator-controller-manager-69bb78d57c-p4l9h\" (UID: \"2fed244b-a135-4d95-a5e5-2eea4bcde7cb\") " pod="metallb-system/metallb-operator-controller-manager-69bb78d57c-p4l9h" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.270383 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2fed244b-a135-4d95-a5e5-2eea4bcde7cb-apiservice-cert\") pod \"metallb-operator-controller-manager-69bb78d57c-p4l9h\" (UID: \"2fed244b-a135-4d95-a5e5-2eea4bcde7cb\") " pod="metallb-system/metallb-operator-controller-manager-69bb78d57c-p4l9h" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.270426 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2fed244b-a135-4d95-a5e5-2eea4bcde7cb-webhook-cert\") pod \"metallb-operator-controller-manager-69bb78d57c-p4l9h\" (UID: \"2fed244b-a135-4d95-a5e5-2eea4bcde7cb\") " pod="metallb-system/metallb-operator-controller-manager-69bb78d57c-p4l9h" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.277799 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2fed244b-a135-4d95-a5e5-2eea4bcde7cb-apiservice-cert\") pod \"metallb-operator-controller-manager-69bb78d57c-p4l9h\" (UID: \"2fed244b-a135-4d95-a5e5-2eea4bcde7cb\") " pod="metallb-system/metallb-operator-controller-manager-69bb78d57c-p4l9h" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.285874 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2fed244b-a135-4d95-a5e5-2eea4bcde7cb-webhook-cert\") pod \"metallb-operator-controller-manager-69bb78d57c-p4l9h\" (UID: \"2fed244b-a135-4d95-a5e5-2eea4bcde7cb\") " pod="metallb-system/metallb-operator-controller-manager-69bb78d57c-p4l9h" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.291874 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz8wf\" (UniqueName: \"kubernetes.io/projected/2fed244b-a135-4d95-a5e5-2eea4bcde7cb-kube-api-access-tz8wf\") pod \"metallb-operator-controller-manager-69bb78d57c-p4l9h\" (UID: \"2fed244b-a135-4d95-a5e5-2eea4bcde7cb\") " pod="metallb-system/metallb-operator-controller-manager-69bb78d57c-p4l9h" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.391644 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-69bb78d57c-p4l9h" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.397193 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw"] Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.398641 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.402271 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.402505 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.410927 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-tkm6t" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.452255 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw"] Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.579143 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bab491e7-4eec-4ba1-975d-fb2468372ade-webhook-cert\") pod \"metallb-operator-webhook-server-5dcb7584cb-dfrlw\" (UID: \"bab491e7-4eec-4ba1-975d-fb2468372ade\") " pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.579662 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctdct\" (UniqueName: \"kubernetes.io/projected/bab491e7-4eec-4ba1-975d-fb2468372ade-kube-api-access-ctdct\") pod \"metallb-operator-webhook-server-5dcb7584cb-dfrlw\" (UID: \"bab491e7-4eec-4ba1-975d-fb2468372ade\") " pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.579801 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bab491e7-4eec-4ba1-975d-fb2468372ade-apiservice-cert\") pod \"metallb-operator-webhook-server-5dcb7584cb-dfrlw\" (UID: \"bab491e7-4eec-4ba1-975d-fb2468372ade\") " pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.681263 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bab491e7-4eec-4ba1-975d-fb2468372ade-webhook-cert\") pod \"metallb-operator-webhook-server-5dcb7584cb-dfrlw\" (UID: \"bab491e7-4eec-4ba1-975d-fb2468372ade\") " pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.681317 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctdct\" (UniqueName: \"kubernetes.io/projected/bab491e7-4eec-4ba1-975d-fb2468372ade-kube-api-access-ctdct\") pod \"metallb-operator-webhook-server-5dcb7584cb-dfrlw\" (UID: \"bab491e7-4eec-4ba1-975d-fb2468372ade\") " pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.681385 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bab491e7-4eec-4ba1-975d-fb2468372ade-apiservice-cert\") pod \"metallb-operator-webhook-server-5dcb7584cb-dfrlw\" (UID: \"bab491e7-4eec-4ba1-975d-fb2468372ade\") " pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.696169 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bab491e7-4eec-4ba1-975d-fb2468372ade-webhook-cert\") pod \"metallb-operator-webhook-server-5dcb7584cb-dfrlw\" (UID: \"bab491e7-4eec-4ba1-975d-fb2468372ade\") " pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.712838 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctdct\" (UniqueName: \"kubernetes.io/projected/bab491e7-4eec-4ba1-975d-fb2468372ade-kube-api-access-ctdct\") pod \"metallb-operator-webhook-server-5dcb7584cb-dfrlw\" (UID: \"bab491e7-4eec-4ba1-975d-fb2468372ade\") " pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.720543 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bab491e7-4eec-4ba1-975d-fb2468372ade-apiservice-cert\") pod \"metallb-operator-webhook-server-5dcb7584cb-dfrlw\" (UID: \"bab491e7-4eec-4ba1-975d-fb2468372ade\") " pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.803266 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" Dec 09 10:03:39 crc kubenswrapper[4824]: I1209 10:03:39.948405 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-69bb78d57c-p4l9h"] Dec 09 10:03:40 crc kubenswrapper[4824]: I1209 10:03:40.336758 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw"] Dec 09 10:03:40 crc kubenswrapper[4824]: I1209 10:03:40.355568 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-69bb78d57c-p4l9h" event={"ID":"2fed244b-a135-4d95-a5e5-2eea4bcde7cb","Type":"ContainerStarted","Data":"c7017e5f042880ca76121f7070eb2eb3f38ede8c2e98fb9cea1c79651ec9a889"} Dec 09 10:03:40 crc kubenswrapper[4824]: W1209 10:03:40.369986 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbab491e7_4eec_4ba1_975d_fb2468372ade.slice/crio-4711fed380705ccfb0d1457cbd6d674f4f2ee506a296f88e1bc2087163a4e82c WatchSource:0}: Error finding container 4711fed380705ccfb0d1457cbd6d674f4f2ee506a296f88e1bc2087163a4e82c: Status 404 returned error can't find the container with id 4711fed380705ccfb0d1457cbd6d674f4f2ee506a296f88e1bc2087163a4e82c Dec 09 10:03:41 crc kubenswrapper[4824]: I1209 10:03:41.368692 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" event={"ID":"bab491e7-4eec-4ba1-975d-fb2468372ade","Type":"ContainerStarted","Data":"4711fed380705ccfb0d1457cbd6d674f4f2ee506a296f88e1bc2087163a4e82c"} Dec 09 10:03:44 crc kubenswrapper[4824]: I1209 10:03:44.404095 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-69bb78d57c-p4l9h" event={"ID":"2fed244b-a135-4d95-a5e5-2eea4bcde7cb","Type":"ContainerStarted","Data":"f3716d785a2a884bb83bd74ab737a8c3134c7c5dbcad6ced4e12e15384406ca5"} Dec 09 10:03:44 crc kubenswrapper[4824]: I1209 10:03:44.405405 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-69bb78d57c-p4l9h" Dec 09 10:03:44 crc kubenswrapper[4824]: I1209 10:03:44.464150 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-69bb78d57c-p4l9h" podStartSLOduration=1.977249311 podStartE2EDuration="5.46412848s" podCreationTimestamp="2025-12-09 10:03:39 +0000 UTC" firstStartedPulling="2025-12-09 10:03:39.957358512 +0000 UTC m=+976.291863179" lastFinishedPulling="2025-12-09 10:03:43.444237681 +0000 UTC m=+979.778742348" observedRunningTime="2025-12-09 10:03:44.434200386 +0000 UTC m=+980.768705083" watchObservedRunningTime="2025-12-09 10:03:44.46412848 +0000 UTC m=+980.798633147" Dec 09 10:03:47 crc kubenswrapper[4824]: I1209 10:03:47.706279 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" event={"ID":"bab491e7-4eec-4ba1-975d-fb2468372ade","Type":"ContainerStarted","Data":"499a81ff89ffe2fe094b2e4d1f08ec2b8f670eadfaddc8a27c804fb362e672a6"} Dec 09 10:03:47 crc kubenswrapper[4824]: I1209 10:03:47.706971 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" Dec 09 10:03:59 crc kubenswrapper[4824]: I1209 10:03:59.810721 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" Dec 09 10:03:59 crc kubenswrapper[4824]: I1209 10:03:59.836639 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" podStartSLOduration=13.875641978 podStartE2EDuration="20.836519627s" podCreationTimestamp="2025-12-09 10:03:39 +0000 UTC" firstStartedPulling="2025-12-09 10:03:40.376289861 +0000 UTC m=+976.710794528" lastFinishedPulling="2025-12-09 10:03:47.33716751 +0000 UTC m=+983.671672177" observedRunningTime="2025-12-09 10:03:47.730478471 +0000 UTC m=+984.064983138" watchObservedRunningTime="2025-12-09 10:03:59.836519627 +0000 UTC m=+996.171024314" Dec 09 10:04:02 crc kubenswrapper[4824]: I1209 10:04:02.861747 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:04:02 crc kubenswrapper[4824]: I1209 10:04:02.862286 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:04:19 crc kubenswrapper[4824]: I1209 10:04:19.395362 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-69bb78d57c-p4l9h" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.323447 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-ntx24"] Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.326642 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.329545 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.330001 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-6spdh" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.330033 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.342628 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-8rvmd"] Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.344287 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-8rvmd" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.345993 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.362689 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-8rvmd"] Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.460658 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-5xfv2"] Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.462316 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-5xfv2" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.464585 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.464752 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.465246 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-s68p2" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.465356 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.469941 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/ad8f524a-8c95-47ab-b74d-9f83331fce76-reloader\") pod \"frr-k8s-ntx24\" (UID: \"ad8f524a-8c95-47ab-b74d-9f83331fce76\") " pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.470479 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/ad8f524a-8c95-47ab-b74d-9f83331fce76-frr-conf\") pod \"frr-k8s-ntx24\" (UID: \"ad8f524a-8c95-47ab-b74d-9f83331fce76\") " pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.470592 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad8f524a-8c95-47ab-b74d-9f83331fce76-metrics-certs\") pod \"frr-k8s-ntx24\" (UID: \"ad8f524a-8c95-47ab-b74d-9f83331fce76\") " pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.470704 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l26rg\" (UniqueName: \"kubernetes.io/projected/bd655e5b-2ac3-4fe8-ad58-6012d0180897-kube-api-access-l26rg\") pod \"frr-k8s-webhook-server-7fcb986d4-8rvmd\" (UID: \"bd655e5b-2ac3-4fe8-ad58-6012d0180897\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-8rvmd" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.470935 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whr8m\" (UniqueName: \"kubernetes.io/projected/ad8f524a-8c95-47ab-b74d-9f83331fce76-kube-api-access-whr8m\") pod \"frr-k8s-ntx24\" (UID: \"ad8f524a-8c95-47ab-b74d-9f83331fce76\") " pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.471068 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/ad8f524a-8c95-47ab-b74d-9f83331fce76-frr-startup\") pod \"frr-k8s-ntx24\" (UID: \"ad8f524a-8c95-47ab-b74d-9f83331fce76\") " pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.471228 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/ad8f524a-8c95-47ab-b74d-9f83331fce76-metrics\") pod \"frr-k8s-ntx24\" (UID: \"ad8f524a-8c95-47ab-b74d-9f83331fce76\") " pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.471375 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bd655e5b-2ac3-4fe8-ad58-6012d0180897-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-8rvmd\" (UID: \"bd655e5b-2ac3-4fe8-ad58-6012d0180897\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-8rvmd" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.471562 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/ad8f524a-8c95-47ab-b74d-9f83331fce76-frr-sockets\") pod \"frr-k8s-ntx24\" (UID: \"ad8f524a-8c95-47ab-b74d-9f83331fce76\") " pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.507653 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-wxgt4"] Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.509060 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-wxgt4" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.520892 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.523743 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-wxgt4"] Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.573355 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/ad8f524a-8c95-47ab-b74d-9f83331fce76-frr-conf\") pod \"frr-k8s-ntx24\" (UID: \"ad8f524a-8c95-47ab-b74d-9f83331fce76\") " pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.573711 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad8f524a-8c95-47ab-b74d-9f83331fce76-metrics-certs\") pod \"frr-k8s-ntx24\" (UID: \"ad8f524a-8c95-47ab-b74d-9f83331fce76\") " pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.573861 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l26rg\" (UniqueName: \"kubernetes.io/projected/bd655e5b-2ac3-4fe8-ad58-6012d0180897-kube-api-access-l26rg\") pod \"frr-k8s-webhook-server-7fcb986d4-8rvmd\" (UID: \"bd655e5b-2ac3-4fe8-ad58-6012d0180897\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-8rvmd" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.573988 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/ad8f524a-8c95-47ab-b74d-9f83331fce76-frr-conf\") pod \"frr-k8s-ntx24\" (UID: \"ad8f524a-8c95-47ab-b74d-9f83331fce76\") " pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.574009 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whr8m\" (UniqueName: \"kubernetes.io/projected/ad8f524a-8c95-47ab-b74d-9f83331fce76-kube-api-access-whr8m\") pod \"frr-k8s-ntx24\" (UID: \"ad8f524a-8c95-47ab-b74d-9f83331fce76\") " pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.574120 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe-metrics-certs\") pod \"speaker-5xfv2\" (UID: \"87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe\") " pod="metallb-system/speaker-5xfv2" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.574166 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k95dd\" (UniqueName: \"kubernetes.io/projected/87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe-kube-api-access-k95dd\") pod \"speaker-5xfv2\" (UID: \"87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe\") " pod="metallb-system/speaker-5xfv2" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.574240 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/ad8f524a-8c95-47ab-b74d-9f83331fce76-frr-startup\") pod \"frr-k8s-ntx24\" (UID: \"ad8f524a-8c95-47ab-b74d-9f83331fce76\") " pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.574324 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/ad8f524a-8c95-47ab-b74d-9f83331fce76-metrics\") pod \"frr-k8s-ntx24\" (UID: \"ad8f524a-8c95-47ab-b74d-9f83331fce76\") " pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.574361 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe-metallb-excludel2\") pod \"speaker-5xfv2\" (UID: \"87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe\") " pod="metallb-system/speaker-5xfv2" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.574422 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bd655e5b-2ac3-4fe8-ad58-6012d0180897-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-8rvmd\" (UID: \"bd655e5b-2ac3-4fe8-ad58-6012d0180897\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-8rvmd" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.574508 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe-memberlist\") pod \"speaker-5xfv2\" (UID: \"87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe\") " pod="metallb-system/speaker-5xfv2" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.574558 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/ad8f524a-8c95-47ab-b74d-9f83331fce76-frr-sockets\") pod \"frr-k8s-ntx24\" (UID: \"ad8f524a-8c95-47ab-b74d-9f83331fce76\") " pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:20 crc kubenswrapper[4824]: E1209 10:04:20.574572 4824 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.574590 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3af18cd9-4124-4f24-83ea-0ad26ad9ae43-metrics-certs\") pod \"controller-f8648f98b-wxgt4\" (UID: \"3af18cd9-4124-4f24-83ea-0ad26ad9ae43\") " pod="metallb-system/controller-f8648f98b-wxgt4" Dec 09 10:04:20 crc kubenswrapper[4824]: E1209 10:04:20.574635 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bd655e5b-2ac3-4fe8-ad58-6012d0180897-cert podName:bd655e5b-2ac3-4fe8-ad58-6012d0180897 nodeName:}" failed. No retries permitted until 2025-12-09 10:04:21.074613645 +0000 UTC m=+1017.409118312 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/bd655e5b-2ac3-4fe8-ad58-6012d0180897-cert") pod "frr-k8s-webhook-server-7fcb986d4-8rvmd" (UID: "bd655e5b-2ac3-4fe8-ad58-6012d0180897") : secret "frr-k8s-webhook-server-cert" not found Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.574664 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3af18cd9-4124-4f24-83ea-0ad26ad9ae43-cert\") pod \"controller-f8648f98b-wxgt4\" (UID: \"3af18cd9-4124-4f24-83ea-0ad26ad9ae43\") " pod="metallb-system/controller-f8648f98b-wxgt4" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.574868 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-586km\" (UniqueName: \"kubernetes.io/projected/3af18cd9-4124-4f24-83ea-0ad26ad9ae43-kube-api-access-586km\") pod \"controller-f8648f98b-wxgt4\" (UID: \"3af18cd9-4124-4f24-83ea-0ad26ad9ae43\") " pod="metallb-system/controller-f8648f98b-wxgt4" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.574914 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/ad8f524a-8c95-47ab-b74d-9f83331fce76-metrics\") pod \"frr-k8s-ntx24\" (UID: \"ad8f524a-8c95-47ab-b74d-9f83331fce76\") " pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.574950 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/ad8f524a-8c95-47ab-b74d-9f83331fce76-reloader\") pod \"frr-k8s-ntx24\" (UID: \"ad8f524a-8c95-47ab-b74d-9f83331fce76\") " pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.574922 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/ad8f524a-8c95-47ab-b74d-9f83331fce76-frr-sockets\") pod \"frr-k8s-ntx24\" (UID: \"ad8f524a-8c95-47ab-b74d-9f83331fce76\") " pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.575245 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/ad8f524a-8c95-47ab-b74d-9f83331fce76-reloader\") pod \"frr-k8s-ntx24\" (UID: \"ad8f524a-8c95-47ab-b74d-9f83331fce76\") " pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.575327 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/ad8f524a-8c95-47ab-b74d-9f83331fce76-frr-startup\") pod \"frr-k8s-ntx24\" (UID: \"ad8f524a-8c95-47ab-b74d-9f83331fce76\") " pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.594649 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad8f524a-8c95-47ab-b74d-9f83331fce76-metrics-certs\") pod \"frr-k8s-ntx24\" (UID: \"ad8f524a-8c95-47ab-b74d-9f83331fce76\") " pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.600662 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l26rg\" (UniqueName: \"kubernetes.io/projected/bd655e5b-2ac3-4fe8-ad58-6012d0180897-kube-api-access-l26rg\") pod \"frr-k8s-webhook-server-7fcb986d4-8rvmd\" (UID: \"bd655e5b-2ac3-4fe8-ad58-6012d0180897\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-8rvmd" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.601149 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whr8m\" (UniqueName: \"kubernetes.io/projected/ad8f524a-8c95-47ab-b74d-9f83331fce76-kube-api-access-whr8m\") pod \"frr-k8s-ntx24\" (UID: \"ad8f524a-8c95-47ab-b74d-9f83331fce76\") " pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.661925 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.678276 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-586km\" (UniqueName: \"kubernetes.io/projected/3af18cd9-4124-4f24-83ea-0ad26ad9ae43-kube-api-access-586km\") pod \"controller-f8648f98b-wxgt4\" (UID: \"3af18cd9-4124-4f24-83ea-0ad26ad9ae43\") " pod="metallb-system/controller-f8648f98b-wxgt4" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.678681 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe-metrics-certs\") pod \"speaker-5xfv2\" (UID: \"87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe\") " pod="metallb-system/speaker-5xfv2" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.678713 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k95dd\" (UniqueName: \"kubernetes.io/projected/87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe-kube-api-access-k95dd\") pod \"speaker-5xfv2\" (UID: \"87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe\") " pod="metallb-system/speaker-5xfv2" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.679614 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe-metallb-excludel2\") pod \"speaker-5xfv2\" (UID: \"87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe\") " pod="metallb-system/speaker-5xfv2" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.679687 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe-memberlist\") pod \"speaker-5xfv2\" (UID: \"87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe\") " pod="metallb-system/speaker-5xfv2" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.679727 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3af18cd9-4124-4f24-83ea-0ad26ad9ae43-metrics-certs\") pod \"controller-f8648f98b-wxgt4\" (UID: \"3af18cd9-4124-4f24-83ea-0ad26ad9ae43\") " pod="metallb-system/controller-f8648f98b-wxgt4" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.679758 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3af18cd9-4124-4f24-83ea-0ad26ad9ae43-cert\") pod \"controller-f8648f98b-wxgt4\" (UID: \"3af18cd9-4124-4f24-83ea-0ad26ad9ae43\") " pod="metallb-system/controller-f8648f98b-wxgt4" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.680485 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe-metallb-excludel2\") pod \"speaker-5xfv2\" (UID: \"87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe\") " pod="metallb-system/speaker-5xfv2" Dec 09 10:04:20 crc kubenswrapper[4824]: E1209 10:04:20.680700 4824 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 09 10:04:20 crc kubenswrapper[4824]: E1209 10:04:20.680885 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe-memberlist podName:87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe nodeName:}" failed. No retries permitted until 2025-12-09 10:04:21.180851787 +0000 UTC m=+1017.515356624 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe-memberlist") pod "speaker-5xfv2" (UID: "87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe") : secret "metallb-memberlist" not found Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.684128 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3af18cd9-4124-4f24-83ea-0ad26ad9ae43-metrics-certs\") pod \"controller-f8648f98b-wxgt4\" (UID: \"3af18cd9-4124-4f24-83ea-0ad26ad9ae43\") " pod="metallb-system/controller-f8648f98b-wxgt4" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.684472 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe-metrics-certs\") pod \"speaker-5xfv2\" (UID: \"87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe\") " pod="metallb-system/speaker-5xfv2" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.684774 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3af18cd9-4124-4f24-83ea-0ad26ad9ae43-cert\") pod \"controller-f8648f98b-wxgt4\" (UID: \"3af18cd9-4124-4f24-83ea-0ad26ad9ae43\") " pod="metallb-system/controller-f8648f98b-wxgt4" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.701326 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k95dd\" (UniqueName: \"kubernetes.io/projected/87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe-kube-api-access-k95dd\") pod \"speaker-5xfv2\" (UID: \"87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe\") " pod="metallb-system/speaker-5xfv2" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.703473 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-586km\" (UniqueName: \"kubernetes.io/projected/3af18cd9-4124-4f24-83ea-0ad26ad9ae43-kube-api-access-586km\") pod \"controller-f8648f98b-wxgt4\" (UID: \"3af18cd9-4124-4f24-83ea-0ad26ad9ae43\") " pod="metallb-system/controller-f8648f98b-wxgt4" Dec 09 10:04:20 crc kubenswrapper[4824]: I1209 10:04:20.832874 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-wxgt4" Dec 09 10:04:21 crc kubenswrapper[4824]: I1209 10:04:21.088251 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bd655e5b-2ac3-4fe8-ad58-6012d0180897-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-8rvmd\" (UID: \"bd655e5b-2ac3-4fe8-ad58-6012d0180897\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-8rvmd" Dec 09 10:04:21 crc kubenswrapper[4824]: I1209 10:04:21.094178 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bd655e5b-2ac3-4fe8-ad58-6012d0180897-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-8rvmd\" (UID: \"bd655e5b-2ac3-4fe8-ad58-6012d0180897\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-8rvmd" Dec 09 10:04:21 crc kubenswrapper[4824]: I1209 10:04:21.194095 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe-memberlist\") pod \"speaker-5xfv2\" (UID: \"87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe\") " pod="metallb-system/speaker-5xfv2" Dec 09 10:04:21 crc kubenswrapper[4824]: E1209 10:04:21.194361 4824 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 09 10:04:21 crc kubenswrapper[4824]: E1209 10:04:21.194433 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe-memberlist podName:87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe nodeName:}" failed. No retries permitted until 2025-12-09 10:04:22.194409421 +0000 UTC m=+1018.528914088 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe-memberlist") pod "speaker-5xfv2" (UID: "87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe") : secret "metallb-memberlist" not found Dec 09 10:04:21 crc kubenswrapper[4824]: I1209 10:04:21.235111 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-wxgt4"] Dec 09 10:04:21 crc kubenswrapper[4824]: I1209 10:04:21.279930 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-8rvmd" Dec 09 10:04:21 crc kubenswrapper[4824]: I1209 10:04:21.807203 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-8rvmd"] Dec 09 10:04:21 crc kubenswrapper[4824]: W1209 10:04:21.813586 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd655e5b_2ac3_4fe8_ad58_6012d0180897.slice/crio-5c41eb97f9bddf6b373f676cccedc4d56f5c2e2282376a5789770a3154993c1f WatchSource:0}: Error finding container 5c41eb97f9bddf6b373f676cccedc4d56f5c2e2282376a5789770a3154993c1f: Status 404 returned error can't find the container with id 5c41eb97f9bddf6b373f676cccedc4d56f5c2e2282376a5789770a3154993c1f Dec 09 10:04:22 crc kubenswrapper[4824]: I1209 10:04:22.024121 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-wxgt4" event={"ID":"3af18cd9-4124-4f24-83ea-0ad26ad9ae43","Type":"ContainerStarted","Data":"e0ef8a33de35123430308c29510bfb498729033ffcc32e69023f03a412dc22cf"} Dec 09 10:04:22 crc kubenswrapper[4824]: I1209 10:04:22.024566 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-wxgt4" event={"ID":"3af18cd9-4124-4f24-83ea-0ad26ad9ae43","Type":"ContainerStarted","Data":"7db44ab4d9712e36348240dbde88e7ceb831da0d9144ade04e056743e6fe40a1"} Dec 09 10:04:22 crc kubenswrapper[4824]: I1209 10:04:22.024582 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-wxgt4" event={"ID":"3af18cd9-4124-4f24-83ea-0ad26ad9ae43","Type":"ContainerStarted","Data":"86963bda138a41c8e66bdcb23c6db49b7cd5d0388effc41e45cb7d2740f11357"} Dec 09 10:04:22 crc kubenswrapper[4824]: I1209 10:04:22.024602 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-wxgt4" Dec 09 10:04:22 crc kubenswrapper[4824]: I1209 10:04:22.025702 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ntx24" event={"ID":"ad8f524a-8c95-47ab-b74d-9f83331fce76","Type":"ContainerStarted","Data":"b75c9986f62f718e4dfd6a617ea23f4e36c4777ecf72100eef71c436ccb0d016"} Dec 09 10:04:22 crc kubenswrapper[4824]: I1209 10:04:22.026746 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-8rvmd" event={"ID":"bd655e5b-2ac3-4fe8-ad58-6012d0180897","Type":"ContainerStarted","Data":"5c41eb97f9bddf6b373f676cccedc4d56f5c2e2282376a5789770a3154993c1f"} Dec 09 10:04:22 crc kubenswrapper[4824]: I1209 10:04:22.076436 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-wxgt4" podStartSLOduration=2.076408219 podStartE2EDuration="2.076408219s" podCreationTimestamp="2025-12-09 10:04:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:04:22.068336995 +0000 UTC m=+1018.402841662" watchObservedRunningTime="2025-12-09 10:04:22.076408219 +0000 UTC m=+1018.410912886" Dec 09 10:04:22 crc kubenswrapper[4824]: I1209 10:04:22.238137 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe-memberlist\") pod \"speaker-5xfv2\" (UID: \"87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe\") " pod="metallb-system/speaker-5xfv2" Dec 09 10:04:22 crc kubenswrapper[4824]: I1209 10:04:22.256707 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe-memberlist\") pod \"speaker-5xfv2\" (UID: \"87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe\") " pod="metallb-system/speaker-5xfv2" Dec 09 10:04:22 crc kubenswrapper[4824]: I1209 10:04:22.282877 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-5xfv2" Dec 09 10:04:22 crc kubenswrapper[4824]: W1209 10:04:22.318339 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87dca6f6_ef3a_4fc2_a4e2_49f5a54257fe.slice/crio-4daa0256e42e525f02b6e7db19f6f29ebffe3ac1324999d0bff07f64e5a5554c WatchSource:0}: Error finding container 4daa0256e42e525f02b6e7db19f6f29ebffe3ac1324999d0bff07f64e5a5554c: Status 404 returned error can't find the container with id 4daa0256e42e525f02b6e7db19f6f29ebffe3ac1324999d0bff07f64e5a5554c Dec 09 10:04:23 crc kubenswrapper[4824]: I1209 10:04:23.036146 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5xfv2" event={"ID":"87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe","Type":"ContainerStarted","Data":"dd1807b856b18c97d0d6ae44f7011055c8dcd704fa540f6f20146e7f61f0b645"} Dec 09 10:04:23 crc kubenswrapper[4824]: I1209 10:04:23.036556 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5xfv2" event={"ID":"87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe","Type":"ContainerStarted","Data":"4daa0256e42e525f02b6e7db19f6f29ebffe3ac1324999d0bff07f64e5a5554c"} Dec 09 10:04:24 crc kubenswrapper[4824]: I1209 10:04:24.077641 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5xfv2" event={"ID":"87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe","Type":"ContainerStarted","Data":"be0007cd96b66472a9c9023391c39579ac217f77267abc5b9fdc06d20c34af47"} Dec 09 10:04:24 crc kubenswrapper[4824]: I1209 10:04:24.078091 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-5xfv2" Dec 09 10:04:24 crc kubenswrapper[4824]: I1209 10:04:24.118741 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-5xfv2" podStartSLOduration=4.118713888 podStartE2EDuration="4.118713888s" podCreationTimestamp="2025-12-09 10:04:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:04:24.109196088 +0000 UTC m=+1020.443700755" watchObservedRunningTime="2025-12-09 10:04:24.118713888 +0000 UTC m=+1020.453218555" Dec 09 10:04:32 crc kubenswrapper[4824]: I1209 10:04:32.291147 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-5xfv2" Dec 09 10:04:32 crc kubenswrapper[4824]: I1209 10:04:32.860923 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:04:32 crc kubenswrapper[4824]: I1209 10:04:32.861365 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:04:33 crc kubenswrapper[4824]: I1209 10:04:33.380834 4824 generic.go:334] "Generic (PLEG): container finished" podID="ad8f524a-8c95-47ab-b74d-9f83331fce76" containerID="50de8be5f556be04d6d6a24d1ea45998a6749b047744c6404b4d0b76e5a060b9" exitCode=0 Dec 09 10:04:33 crc kubenswrapper[4824]: I1209 10:04:33.380945 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ntx24" event={"ID":"ad8f524a-8c95-47ab-b74d-9f83331fce76","Type":"ContainerDied","Data":"50de8be5f556be04d6d6a24d1ea45998a6749b047744c6404b4d0b76e5a060b9"} Dec 09 10:04:33 crc kubenswrapper[4824]: I1209 10:04:33.382939 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-8rvmd" event={"ID":"bd655e5b-2ac3-4fe8-ad58-6012d0180897","Type":"ContainerStarted","Data":"bb7e099dcce79c54eefc8387db330fcdbf2020412f6b56917e96a81d1558019b"} Dec 09 10:04:33 crc kubenswrapper[4824]: I1209 10:04:33.383077 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-8rvmd" Dec 09 10:04:33 crc kubenswrapper[4824]: I1209 10:04:33.426324 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-8rvmd" podStartSLOduration=2.948657801 podStartE2EDuration="13.426297241s" podCreationTimestamp="2025-12-09 10:04:20 +0000 UTC" firstStartedPulling="2025-12-09 10:04:21.8181179 +0000 UTC m=+1018.152622567" lastFinishedPulling="2025-12-09 10:04:32.29575734 +0000 UTC m=+1028.630262007" observedRunningTime="2025-12-09 10:04:33.421087316 +0000 UTC m=+1029.755591983" watchObservedRunningTime="2025-12-09 10:04:33.426297241 +0000 UTC m=+1029.760801908" Dec 09 10:04:34 crc kubenswrapper[4824]: I1209 10:04:34.394105 4824 generic.go:334] "Generic (PLEG): container finished" podID="ad8f524a-8c95-47ab-b74d-9f83331fce76" containerID="fc1b5a48b02f727cc3877c97adfa99916b035cef995bc11d346aa5c97d43a91a" exitCode=0 Dec 09 10:04:34 crc kubenswrapper[4824]: I1209 10:04:34.394204 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ntx24" event={"ID":"ad8f524a-8c95-47ab-b74d-9f83331fce76","Type":"ContainerDied","Data":"fc1b5a48b02f727cc3877c97adfa99916b035cef995bc11d346aa5c97d43a91a"} Dec 09 10:04:35 crc kubenswrapper[4824]: I1209 10:04:35.405230 4824 generic.go:334] "Generic (PLEG): container finished" podID="ad8f524a-8c95-47ab-b74d-9f83331fce76" containerID="ff938e533348d449c87c54e1a042f757fa13c0d8bd758f0e8ee207c62b611017" exitCode=0 Dec 09 10:04:35 crc kubenswrapper[4824]: I1209 10:04:35.405296 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ntx24" event={"ID":"ad8f524a-8c95-47ab-b74d-9f83331fce76","Type":"ContainerDied","Data":"ff938e533348d449c87c54e1a042f757fa13c0d8bd758f0e8ee207c62b611017"} Dec 09 10:04:35 crc kubenswrapper[4824]: I1209 10:04:35.847983 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-qkl57"] Dec 09 10:04:35 crc kubenswrapper[4824]: I1209 10:04:35.849875 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qkl57" Dec 09 10:04:35 crc kubenswrapper[4824]: I1209 10:04:35.853312 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 09 10:04:35 crc kubenswrapper[4824]: I1209 10:04:35.853341 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-p5dzv" Dec 09 10:04:35 crc kubenswrapper[4824]: I1209 10:04:35.853888 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 09 10:04:35 crc kubenswrapper[4824]: I1209 10:04:35.876810 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-qkl57"] Dec 09 10:04:35 crc kubenswrapper[4824]: I1209 10:04:35.887927 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb2w8\" (UniqueName: \"kubernetes.io/projected/08eee4f9-dee4-4c93-af12-5911838ef5dd-kube-api-access-rb2w8\") pod \"openstack-operator-index-qkl57\" (UID: \"08eee4f9-dee4-4c93-af12-5911838ef5dd\") " pod="openstack-operators/openstack-operator-index-qkl57" Dec 09 10:04:35 crc kubenswrapper[4824]: I1209 10:04:35.990213 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb2w8\" (UniqueName: \"kubernetes.io/projected/08eee4f9-dee4-4c93-af12-5911838ef5dd-kube-api-access-rb2w8\") pod \"openstack-operator-index-qkl57\" (UID: \"08eee4f9-dee4-4c93-af12-5911838ef5dd\") " pod="openstack-operators/openstack-operator-index-qkl57" Dec 09 10:04:36 crc kubenswrapper[4824]: I1209 10:04:36.020968 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb2w8\" (UniqueName: \"kubernetes.io/projected/08eee4f9-dee4-4c93-af12-5911838ef5dd-kube-api-access-rb2w8\") pod \"openstack-operator-index-qkl57\" (UID: \"08eee4f9-dee4-4c93-af12-5911838ef5dd\") " pod="openstack-operators/openstack-operator-index-qkl57" Dec 09 10:04:36 crc kubenswrapper[4824]: I1209 10:04:36.185921 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qkl57" Dec 09 10:04:36 crc kubenswrapper[4824]: I1209 10:04:36.435165 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ntx24" event={"ID":"ad8f524a-8c95-47ab-b74d-9f83331fce76","Type":"ContainerStarted","Data":"5896f9eb33a72759d7de6ee9bed25634ccf1414f823ac094e6ea384bd9101d6f"} Dec 09 10:04:36 crc kubenswrapper[4824]: I1209 10:04:36.435531 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ntx24" event={"ID":"ad8f524a-8c95-47ab-b74d-9f83331fce76","Type":"ContainerStarted","Data":"8b2276c4909df36d50f430093ec7a9d213e3472040c3d04dc67f897a7081d3bc"} Dec 09 10:04:36 crc kubenswrapper[4824]: I1209 10:04:36.435542 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ntx24" event={"ID":"ad8f524a-8c95-47ab-b74d-9f83331fce76","Type":"ContainerStarted","Data":"7123ec58c0ff998d60e21c7177b2ee09325340ea626ba5ccbef31eecc03dd982"} Dec 09 10:04:36 crc kubenswrapper[4824]: I1209 10:04:36.435552 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ntx24" event={"ID":"ad8f524a-8c95-47ab-b74d-9f83331fce76","Type":"ContainerStarted","Data":"923c4295cf50c2c589b27821064920e6ef72cb71891a3322348ad6ad4d10e10f"} Dec 09 10:04:36 crc kubenswrapper[4824]: I1209 10:04:36.831354 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-qkl57"] Dec 09 10:04:37 crc kubenswrapper[4824]: I1209 10:04:37.445596 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qkl57" event={"ID":"08eee4f9-dee4-4c93-af12-5911838ef5dd","Type":"ContainerStarted","Data":"987437197bc9e7f0165487a60609383b7fd2ad0b1677c958e9b126f7840efe63"} Dec 09 10:04:37 crc kubenswrapper[4824]: I1209 10:04:37.449622 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ntx24" event={"ID":"ad8f524a-8c95-47ab-b74d-9f83331fce76","Type":"ContainerStarted","Data":"db927dbb4670d706ea865bad8dbbb2d1896a40959bad08f953f751703168664b"} Dec 09 10:04:37 crc kubenswrapper[4824]: I1209 10:04:37.449676 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ntx24" event={"ID":"ad8f524a-8c95-47ab-b74d-9f83331fce76","Type":"ContainerStarted","Data":"1cebb62b79387292f44617976359e0443b6d96ae187a7f53a10dcccaa6ebdbfa"} Dec 09 10:04:37 crc kubenswrapper[4824]: I1209 10:04:37.449941 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:37 crc kubenswrapper[4824]: I1209 10:04:37.482401 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-ntx24" podStartSLOduration=6.679238528 podStartE2EDuration="17.482380479s" podCreationTimestamp="2025-12-09 10:04:20 +0000 UTC" firstStartedPulling="2025-12-09 10:04:21.498465644 +0000 UTC m=+1017.832970311" lastFinishedPulling="2025-12-09 10:04:32.301607595 +0000 UTC m=+1028.636112262" observedRunningTime="2025-12-09 10:04:37.475058188 +0000 UTC m=+1033.809562855" watchObservedRunningTime="2025-12-09 10:04:37.482380479 +0000 UTC m=+1033.816885146" Dec 09 10:04:39 crc kubenswrapper[4824]: I1209 10:04:39.218797 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-qkl57"] Dec 09 10:04:39 crc kubenswrapper[4824]: I1209 10:04:39.828831 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-2jnrg"] Dec 09 10:04:39 crc kubenswrapper[4824]: I1209 10:04:39.830913 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2jnrg" Dec 09 10:04:39 crc kubenswrapper[4824]: I1209 10:04:39.849737 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2jnrg"] Dec 09 10:04:39 crc kubenswrapper[4824]: I1209 10:04:39.990599 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8sj9\" (UniqueName: \"kubernetes.io/projected/25c6995d-6a34-43fc-9b1b-6c34ed9aec46-kube-api-access-k8sj9\") pod \"openstack-operator-index-2jnrg\" (UID: \"25c6995d-6a34-43fc-9b1b-6c34ed9aec46\") " pod="openstack-operators/openstack-operator-index-2jnrg" Dec 09 10:04:40 crc kubenswrapper[4824]: I1209 10:04:40.093119 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8sj9\" (UniqueName: \"kubernetes.io/projected/25c6995d-6a34-43fc-9b1b-6c34ed9aec46-kube-api-access-k8sj9\") pod \"openstack-operator-index-2jnrg\" (UID: \"25c6995d-6a34-43fc-9b1b-6c34ed9aec46\") " pod="openstack-operators/openstack-operator-index-2jnrg" Dec 09 10:04:40 crc kubenswrapper[4824]: I1209 10:04:40.120899 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8sj9\" (UniqueName: \"kubernetes.io/projected/25c6995d-6a34-43fc-9b1b-6c34ed9aec46-kube-api-access-k8sj9\") pod \"openstack-operator-index-2jnrg\" (UID: \"25c6995d-6a34-43fc-9b1b-6c34ed9aec46\") " pod="openstack-operators/openstack-operator-index-2jnrg" Dec 09 10:04:40 crc kubenswrapper[4824]: I1209 10:04:40.179551 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2jnrg" Dec 09 10:04:40 crc kubenswrapper[4824]: I1209 10:04:40.489205 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qkl57" event={"ID":"08eee4f9-dee4-4c93-af12-5911838ef5dd","Type":"ContainerStarted","Data":"27bab33f44a76cdd6f8a97c151276ca8a7dcd330a6c20b5328c1c0d9d4a99be8"} Dec 09 10:04:40 crc kubenswrapper[4824]: I1209 10:04:40.489319 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-qkl57" podUID="08eee4f9-dee4-4c93-af12-5911838ef5dd" containerName="registry-server" containerID="cri-o://27bab33f44a76cdd6f8a97c151276ca8a7dcd330a6c20b5328c1c0d9d4a99be8" gracePeriod=2 Dec 09 10:04:40 crc kubenswrapper[4824]: I1209 10:04:40.510912 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-qkl57" podStartSLOduration=2.720935956 podStartE2EDuration="5.510887935s" podCreationTimestamp="2025-12-09 10:04:35 +0000 UTC" firstStartedPulling="2025-12-09 10:04:36.840998092 +0000 UTC m=+1033.175502759" lastFinishedPulling="2025-12-09 10:04:39.630950071 +0000 UTC m=+1035.965454738" observedRunningTime="2025-12-09 10:04:40.510251204 +0000 UTC m=+1036.844755891" watchObservedRunningTime="2025-12-09 10:04:40.510887935 +0000 UTC m=+1036.845392602" Dec 09 10:04:40 crc kubenswrapper[4824]: I1209 10:04:40.646999 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2jnrg"] Dec 09 10:04:40 crc kubenswrapper[4824]: W1209 10:04:40.656969 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25c6995d_6a34_43fc_9b1b_6c34ed9aec46.slice/crio-ad2ef0749a4ca9869e85b9ea0351646b14f8fb9e2708b8b8a20cd617d3b53805 WatchSource:0}: Error finding container ad2ef0749a4ca9869e85b9ea0351646b14f8fb9e2708b8b8a20cd617d3b53805: Status 404 returned error can't find the container with id ad2ef0749a4ca9869e85b9ea0351646b14f8fb9e2708b8b8a20cd617d3b53805 Dec 09 10:04:40 crc kubenswrapper[4824]: I1209 10:04:40.663265 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:40 crc kubenswrapper[4824]: I1209 10:04:40.703393 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:40 crc kubenswrapper[4824]: I1209 10:04:40.837964 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-wxgt4" Dec 09 10:04:41 crc kubenswrapper[4824]: I1209 10:04:41.465729 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qkl57" Dec 09 10:04:41 crc kubenswrapper[4824]: I1209 10:04:41.511979 4824 generic.go:334] "Generic (PLEG): container finished" podID="08eee4f9-dee4-4c93-af12-5911838ef5dd" containerID="27bab33f44a76cdd6f8a97c151276ca8a7dcd330a6c20b5328c1c0d9d4a99be8" exitCode=0 Dec 09 10:04:41 crc kubenswrapper[4824]: I1209 10:04:41.512074 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qkl57" Dec 09 10:04:41 crc kubenswrapper[4824]: I1209 10:04:41.512080 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qkl57" event={"ID":"08eee4f9-dee4-4c93-af12-5911838ef5dd","Type":"ContainerDied","Data":"27bab33f44a76cdd6f8a97c151276ca8a7dcd330a6c20b5328c1c0d9d4a99be8"} Dec 09 10:04:41 crc kubenswrapper[4824]: I1209 10:04:41.512279 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qkl57" event={"ID":"08eee4f9-dee4-4c93-af12-5911838ef5dd","Type":"ContainerDied","Data":"987437197bc9e7f0165487a60609383b7fd2ad0b1677c958e9b126f7840efe63"} Dec 09 10:04:41 crc kubenswrapper[4824]: I1209 10:04:41.512310 4824 scope.go:117] "RemoveContainer" containerID="27bab33f44a76cdd6f8a97c151276ca8a7dcd330a6c20b5328c1c0d9d4a99be8" Dec 09 10:04:41 crc kubenswrapper[4824]: I1209 10:04:41.515398 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2jnrg" event={"ID":"25c6995d-6a34-43fc-9b1b-6c34ed9aec46","Type":"ContainerStarted","Data":"a314e961a2fcdf2c002cb7e7e4673199d607efd9a9dae356d4aad8f2adbaf060"} Dec 09 10:04:41 crc kubenswrapper[4824]: I1209 10:04:41.515451 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2jnrg" event={"ID":"25c6995d-6a34-43fc-9b1b-6c34ed9aec46","Type":"ContainerStarted","Data":"ad2ef0749a4ca9869e85b9ea0351646b14f8fb9e2708b8b8a20cd617d3b53805"} Dec 09 10:04:41 crc kubenswrapper[4824]: I1209 10:04:41.523579 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rb2w8\" (UniqueName: \"kubernetes.io/projected/08eee4f9-dee4-4c93-af12-5911838ef5dd-kube-api-access-rb2w8\") pod \"08eee4f9-dee4-4c93-af12-5911838ef5dd\" (UID: \"08eee4f9-dee4-4c93-af12-5911838ef5dd\") " Dec 09 10:04:41 crc kubenswrapper[4824]: I1209 10:04:41.531665 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08eee4f9-dee4-4c93-af12-5911838ef5dd-kube-api-access-rb2w8" (OuterVolumeSpecName: "kube-api-access-rb2w8") pod "08eee4f9-dee4-4c93-af12-5911838ef5dd" (UID: "08eee4f9-dee4-4c93-af12-5911838ef5dd"). InnerVolumeSpecName "kube-api-access-rb2w8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:04:41 crc kubenswrapper[4824]: I1209 10:04:41.549138 4824 scope.go:117] "RemoveContainer" containerID="27bab33f44a76cdd6f8a97c151276ca8a7dcd330a6c20b5328c1c0d9d4a99be8" Dec 09 10:04:41 crc kubenswrapper[4824]: E1209 10:04:41.552920 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27bab33f44a76cdd6f8a97c151276ca8a7dcd330a6c20b5328c1c0d9d4a99be8\": container with ID starting with 27bab33f44a76cdd6f8a97c151276ca8a7dcd330a6c20b5328c1c0d9d4a99be8 not found: ID does not exist" containerID="27bab33f44a76cdd6f8a97c151276ca8a7dcd330a6c20b5328c1c0d9d4a99be8" Dec 09 10:04:41 crc kubenswrapper[4824]: I1209 10:04:41.552996 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27bab33f44a76cdd6f8a97c151276ca8a7dcd330a6c20b5328c1c0d9d4a99be8"} err="failed to get container status \"27bab33f44a76cdd6f8a97c151276ca8a7dcd330a6c20b5328c1c0d9d4a99be8\": rpc error: code = NotFound desc = could not find container \"27bab33f44a76cdd6f8a97c151276ca8a7dcd330a6c20b5328c1c0d9d4a99be8\": container with ID starting with 27bab33f44a76cdd6f8a97c151276ca8a7dcd330a6c20b5328c1c0d9d4a99be8 not found: ID does not exist" Dec 09 10:04:41 crc kubenswrapper[4824]: I1209 10:04:41.556189 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-2jnrg" podStartSLOduration=2.482072386 podStartE2EDuration="2.556162304s" podCreationTimestamp="2025-12-09 10:04:39 +0000 UTC" firstStartedPulling="2025-12-09 10:04:40.661310041 +0000 UTC m=+1036.995814708" lastFinishedPulling="2025-12-09 10:04:40.735399959 +0000 UTC m=+1037.069904626" observedRunningTime="2025-12-09 10:04:41.546048216 +0000 UTC m=+1037.880552903" watchObservedRunningTime="2025-12-09 10:04:41.556162304 +0000 UTC m=+1037.890666972" Dec 09 10:04:41 crc kubenswrapper[4824]: I1209 10:04:41.624652 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rb2w8\" (UniqueName: \"kubernetes.io/projected/08eee4f9-dee4-4c93-af12-5911838ef5dd-kube-api-access-rb2w8\") on node \"crc\" DevicePath \"\"" Dec 09 10:04:41 crc kubenswrapper[4824]: I1209 10:04:41.842496 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-qkl57"] Dec 09 10:04:41 crc kubenswrapper[4824]: I1209 10:04:41.848504 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-qkl57"] Dec 09 10:04:41 crc kubenswrapper[4824]: I1209 10:04:41.924363 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08eee4f9-dee4-4c93-af12-5911838ef5dd" path="/var/lib/kubelet/pods/08eee4f9-dee4-4c93-af12-5911838ef5dd/volumes" Dec 09 10:04:50 crc kubenswrapper[4824]: I1209 10:04:50.180114 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-2jnrg" Dec 09 10:04:50 crc kubenswrapper[4824]: I1209 10:04:50.180913 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-2jnrg" Dec 09 10:04:50 crc kubenswrapper[4824]: I1209 10:04:50.215474 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-2jnrg" Dec 09 10:04:50 crc kubenswrapper[4824]: I1209 10:04:50.614829 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-2jnrg" Dec 09 10:04:50 crc kubenswrapper[4824]: I1209 10:04:50.666802 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-ntx24" Dec 09 10:04:51 crc kubenswrapper[4824]: I1209 10:04:51.290216 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-8rvmd" Dec 09 10:05:02 crc kubenswrapper[4824]: I1209 10:05:02.860700 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:05:02 crc kubenswrapper[4824]: I1209 10:05:02.861330 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:05:02 crc kubenswrapper[4824]: I1209 10:05:02.861375 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 10:05:02 crc kubenswrapper[4824]: I1209 10:05:02.862118 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5abd964d470ad0e2def92ff94c5fb5265d13f56cfeabd66524120ead526f4c21"} pod="openshift-machine-config-operator/machine-config-daemon-dth8x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 10:05:02 crc kubenswrapper[4824]: I1209 10:05:02.862228 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" containerID="cri-o://5abd964d470ad0e2def92ff94c5fb5265d13f56cfeabd66524120ead526f4c21" gracePeriod=600 Dec 09 10:05:03 crc kubenswrapper[4824]: I1209 10:05:03.723852 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerID="5abd964d470ad0e2def92ff94c5fb5265d13f56cfeabd66524120ead526f4c21" exitCode=0 Dec 09 10:05:03 crc kubenswrapper[4824]: I1209 10:05:03.723930 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerDied","Data":"5abd964d470ad0e2def92ff94c5fb5265d13f56cfeabd66524120ead526f4c21"} Dec 09 10:05:03 crc kubenswrapper[4824]: I1209 10:05:03.724319 4824 scope.go:117] "RemoveContainer" containerID="ef93ca94e26d32583c5e2e8c602aa45c24cbdf5e3b1e4f117c890d24188f2377" Dec 09 10:05:04 crc kubenswrapper[4824]: I1209 10:05:04.737503 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerStarted","Data":"b01bf8ba5eb26e2cf11005bb5cc48bd36d1aa75d59e84d3c71f092c10deef9e7"} Dec 09 10:05:12 crc kubenswrapper[4824]: I1209 10:05:12.478144 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk"] Dec 09 10:05:12 crc kubenswrapper[4824]: E1209 10:05:12.479290 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08eee4f9-dee4-4c93-af12-5911838ef5dd" containerName="registry-server" Dec 09 10:05:12 crc kubenswrapper[4824]: I1209 10:05:12.479306 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="08eee4f9-dee4-4c93-af12-5911838ef5dd" containerName="registry-server" Dec 09 10:05:12 crc kubenswrapper[4824]: I1209 10:05:12.479525 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="08eee4f9-dee4-4c93-af12-5911838ef5dd" containerName="registry-server" Dec 09 10:05:12 crc kubenswrapper[4824]: I1209 10:05:12.480897 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk" Dec 09 10:05:12 crc kubenswrapper[4824]: I1209 10:05:12.484513 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-w9b54" Dec 09 10:05:12 crc kubenswrapper[4824]: I1209 10:05:12.499089 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk"] Dec 09 10:05:12 crc kubenswrapper[4824]: I1209 10:05:12.537571 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6910cb67-1871-410a-bc4e-91289bed7d5a-util\") pod \"d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk\" (UID: \"6910cb67-1871-410a-bc4e-91289bed7d5a\") " pod="openstack-operators/d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk" Dec 09 10:05:12 crc kubenswrapper[4824]: I1209 10:05:12.537639 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6910cb67-1871-410a-bc4e-91289bed7d5a-bundle\") pod \"d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk\" (UID: \"6910cb67-1871-410a-bc4e-91289bed7d5a\") " pod="openstack-operators/d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk" Dec 09 10:05:12 crc kubenswrapper[4824]: I1209 10:05:12.537715 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f2lz\" (UniqueName: \"kubernetes.io/projected/6910cb67-1871-410a-bc4e-91289bed7d5a-kube-api-access-6f2lz\") pod \"d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk\" (UID: \"6910cb67-1871-410a-bc4e-91289bed7d5a\") " pod="openstack-operators/d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk" Dec 09 10:05:12 crc kubenswrapper[4824]: I1209 10:05:12.639315 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6910cb67-1871-410a-bc4e-91289bed7d5a-util\") pod \"d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk\" (UID: \"6910cb67-1871-410a-bc4e-91289bed7d5a\") " pod="openstack-operators/d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk" Dec 09 10:05:12 crc kubenswrapper[4824]: I1209 10:05:12.639380 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6910cb67-1871-410a-bc4e-91289bed7d5a-bundle\") pod \"d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk\" (UID: \"6910cb67-1871-410a-bc4e-91289bed7d5a\") " pod="openstack-operators/d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk" Dec 09 10:05:12 crc kubenswrapper[4824]: I1209 10:05:12.639431 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f2lz\" (UniqueName: \"kubernetes.io/projected/6910cb67-1871-410a-bc4e-91289bed7d5a-kube-api-access-6f2lz\") pod \"d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk\" (UID: \"6910cb67-1871-410a-bc4e-91289bed7d5a\") " pod="openstack-operators/d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk" Dec 09 10:05:12 crc kubenswrapper[4824]: I1209 10:05:12.640054 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6910cb67-1871-410a-bc4e-91289bed7d5a-util\") pod \"d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk\" (UID: \"6910cb67-1871-410a-bc4e-91289bed7d5a\") " pod="openstack-operators/d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk" Dec 09 10:05:12 crc kubenswrapper[4824]: I1209 10:05:12.640548 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6910cb67-1871-410a-bc4e-91289bed7d5a-bundle\") pod \"d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk\" (UID: \"6910cb67-1871-410a-bc4e-91289bed7d5a\") " pod="openstack-operators/d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk" Dec 09 10:05:12 crc kubenswrapper[4824]: I1209 10:05:12.666596 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f2lz\" (UniqueName: \"kubernetes.io/projected/6910cb67-1871-410a-bc4e-91289bed7d5a-kube-api-access-6f2lz\") pod \"d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk\" (UID: \"6910cb67-1871-410a-bc4e-91289bed7d5a\") " pod="openstack-operators/d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk" Dec 09 10:05:12 crc kubenswrapper[4824]: I1209 10:05:12.812216 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk" Dec 09 10:05:13 crc kubenswrapper[4824]: I1209 10:05:13.484851 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk"] Dec 09 10:05:13 crc kubenswrapper[4824]: I1209 10:05:13.978131 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk" event={"ID":"6910cb67-1871-410a-bc4e-91289bed7d5a","Type":"ContainerStarted","Data":"f271778151987d44dd5af2a3353db916b74c5b5bf2135519da9046253db8c2b5"} Dec 09 10:05:13 crc kubenswrapper[4824]: I1209 10:05:13.978505 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk" event={"ID":"6910cb67-1871-410a-bc4e-91289bed7d5a","Type":"ContainerStarted","Data":"e8cc2108a815e60d1cfbbb864fb2e94de83829c7d104830d41b87e941de2a376"} Dec 09 10:05:15 crc kubenswrapper[4824]: I1209 10:05:15.008861 4824 generic.go:334] "Generic (PLEG): container finished" podID="6910cb67-1871-410a-bc4e-91289bed7d5a" containerID="f271778151987d44dd5af2a3353db916b74c5b5bf2135519da9046253db8c2b5" exitCode=0 Dec 09 10:05:15 crc kubenswrapper[4824]: I1209 10:05:15.009271 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk" event={"ID":"6910cb67-1871-410a-bc4e-91289bed7d5a","Type":"ContainerDied","Data":"f271778151987d44dd5af2a3353db916b74c5b5bf2135519da9046253db8c2b5"} Dec 09 10:05:16 crc kubenswrapper[4824]: I1209 10:05:16.316491 4824 generic.go:334] "Generic (PLEG): container finished" podID="6910cb67-1871-410a-bc4e-91289bed7d5a" containerID="88373fdc93211485db12926b59e53cdc7667c2823625927d35bf022cb490e936" exitCode=0 Dec 09 10:05:16 crc kubenswrapper[4824]: I1209 10:05:16.317507 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk" event={"ID":"6910cb67-1871-410a-bc4e-91289bed7d5a","Type":"ContainerDied","Data":"88373fdc93211485db12926b59e53cdc7667c2823625927d35bf022cb490e936"} Dec 09 10:05:17 crc kubenswrapper[4824]: I1209 10:05:17.328268 4824 generic.go:334] "Generic (PLEG): container finished" podID="6910cb67-1871-410a-bc4e-91289bed7d5a" containerID="37c676a4d8f3fb22ecf15e01decd4b9e99d9bff8ee4b15b40452ac68e93a0ce3" exitCode=0 Dec 09 10:05:17 crc kubenswrapper[4824]: I1209 10:05:17.328315 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk" event={"ID":"6910cb67-1871-410a-bc4e-91289bed7d5a","Type":"ContainerDied","Data":"37c676a4d8f3fb22ecf15e01decd4b9e99d9bff8ee4b15b40452ac68e93a0ce3"} Dec 09 10:05:18 crc kubenswrapper[4824]: I1209 10:05:18.689226 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk" Dec 09 10:05:18 crc kubenswrapper[4824]: I1209 10:05:18.862878 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6f2lz\" (UniqueName: \"kubernetes.io/projected/6910cb67-1871-410a-bc4e-91289bed7d5a-kube-api-access-6f2lz\") pod \"6910cb67-1871-410a-bc4e-91289bed7d5a\" (UID: \"6910cb67-1871-410a-bc4e-91289bed7d5a\") " Dec 09 10:05:18 crc kubenswrapper[4824]: I1209 10:05:18.862950 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6910cb67-1871-410a-bc4e-91289bed7d5a-bundle\") pod \"6910cb67-1871-410a-bc4e-91289bed7d5a\" (UID: \"6910cb67-1871-410a-bc4e-91289bed7d5a\") " Dec 09 10:05:18 crc kubenswrapper[4824]: I1209 10:05:18.863158 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6910cb67-1871-410a-bc4e-91289bed7d5a-util\") pod \"6910cb67-1871-410a-bc4e-91289bed7d5a\" (UID: \"6910cb67-1871-410a-bc4e-91289bed7d5a\") " Dec 09 10:05:18 crc kubenswrapper[4824]: I1209 10:05:18.864085 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6910cb67-1871-410a-bc4e-91289bed7d5a-bundle" (OuterVolumeSpecName: "bundle") pod "6910cb67-1871-410a-bc4e-91289bed7d5a" (UID: "6910cb67-1871-410a-bc4e-91289bed7d5a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:05:18 crc kubenswrapper[4824]: I1209 10:05:18.871227 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6910cb67-1871-410a-bc4e-91289bed7d5a-kube-api-access-6f2lz" (OuterVolumeSpecName: "kube-api-access-6f2lz") pod "6910cb67-1871-410a-bc4e-91289bed7d5a" (UID: "6910cb67-1871-410a-bc4e-91289bed7d5a"). InnerVolumeSpecName "kube-api-access-6f2lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:05:18 crc kubenswrapper[4824]: I1209 10:05:18.890929 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6910cb67-1871-410a-bc4e-91289bed7d5a-util" (OuterVolumeSpecName: "util") pod "6910cb67-1871-410a-bc4e-91289bed7d5a" (UID: "6910cb67-1871-410a-bc4e-91289bed7d5a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:05:18 crc kubenswrapper[4824]: I1209 10:05:18.965367 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6f2lz\" (UniqueName: \"kubernetes.io/projected/6910cb67-1871-410a-bc4e-91289bed7d5a-kube-api-access-6f2lz\") on node \"crc\" DevicePath \"\"" Dec 09 10:05:18 crc kubenswrapper[4824]: I1209 10:05:18.965412 4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6910cb67-1871-410a-bc4e-91289bed7d5a-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:05:18 crc kubenswrapper[4824]: I1209 10:05:18.965424 4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6910cb67-1871-410a-bc4e-91289bed7d5a-util\") on node \"crc\" DevicePath \"\"" Dec 09 10:05:19 crc kubenswrapper[4824]: I1209 10:05:19.354527 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk" event={"ID":"6910cb67-1871-410a-bc4e-91289bed7d5a","Type":"ContainerDied","Data":"e8cc2108a815e60d1cfbbb864fb2e94de83829c7d104830d41b87e941de2a376"} Dec 09 10:05:19 crc kubenswrapper[4824]: I1209 10:05:19.354583 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8cc2108a815e60d1cfbbb864fb2e94de83829c7d104830d41b87e941de2a376" Dec 09 10:05:19 crc kubenswrapper[4824]: I1209 10:05:19.354663 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk" Dec 09 10:05:27 crc kubenswrapper[4824]: I1209 10:05:27.977970 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-b846b8d48-dhtnn"] Dec 09 10:05:27 crc kubenswrapper[4824]: E1209 10:05:27.978947 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6910cb67-1871-410a-bc4e-91289bed7d5a" containerName="extract" Dec 09 10:05:27 crc kubenswrapper[4824]: I1209 10:05:27.978964 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6910cb67-1871-410a-bc4e-91289bed7d5a" containerName="extract" Dec 09 10:05:27 crc kubenswrapper[4824]: E1209 10:05:27.979009 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6910cb67-1871-410a-bc4e-91289bed7d5a" containerName="pull" Dec 09 10:05:27 crc kubenswrapper[4824]: I1209 10:05:27.979022 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6910cb67-1871-410a-bc4e-91289bed7d5a" containerName="pull" Dec 09 10:05:27 crc kubenswrapper[4824]: E1209 10:05:27.979037 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6910cb67-1871-410a-bc4e-91289bed7d5a" containerName="util" Dec 09 10:05:27 crc kubenswrapper[4824]: I1209 10:05:27.979043 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6910cb67-1871-410a-bc4e-91289bed7d5a" containerName="util" Dec 09 10:05:27 crc kubenswrapper[4824]: I1209 10:05:27.979219 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6910cb67-1871-410a-bc4e-91289bed7d5a" containerName="extract" Dec 09 10:05:27 crc kubenswrapper[4824]: I1209 10:05:27.980366 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-b846b8d48-dhtnn" Dec 09 10:05:27 crc kubenswrapper[4824]: I1209 10:05:27.991438 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-dd9rd" Dec 09 10:05:28 crc kubenswrapper[4824]: I1209 10:05:28.012704 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-b846b8d48-dhtnn"] Dec 09 10:05:28 crc kubenswrapper[4824]: I1209 10:05:28.065631 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsqt9\" (UniqueName: \"kubernetes.io/projected/ebb1c8e1-3960-4c38-969b-c483143c87aa-kube-api-access-fsqt9\") pod \"openstack-operator-controller-operator-b846b8d48-dhtnn\" (UID: \"ebb1c8e1-3960-4c38-969b-c483143c87aa\") " pod="openstack-operators/openstack-operator-controller-operator-b846b8d48-dhtnn" Dec 09 10:05:28 crc kubenswrapper[4824]: I1209 10:05:28.167602 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsqt9\" (UniqueName: \"kubernetes.io/projected/ebb1c8e1-3960-4c38-969b-c483143c87aa-kube-api-access-fsqt9\") pod \"openstack-operator-controller-operator-b846b8d48-dhtnn\" (UID: \"ebb1c8e1-3960-4c38-969b-c483143c87aa\") " pod="openstack-operators/openstack-operator-controller-operator-b846b8d48-dhtnn" Dec 09 10:05:28 crc kubenswrapper[4824]: I1209 10:05:28.191150 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsqt9\" (UniqueName: \"kubernetes.io/projected/ebb1c8e1-3960-4c38-969b-c483143c87aa-kube-api-access-fsqt9\") pod \"openstack-operator-controller-operator-b846b8d48-dhtnn\" (UID: \"ebb1c8e1-3960-4c38-969b-c483143c87aa\") " pod="openstack-operators/openstack-operator-controller-operator-b846b8d48-dhtnn" Dec 09 10:05:28 crc kubenswrapper[4824]: I1209 10:05:28.367509 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-b846b8d48-dhtnn" Dec 09 10:05:29 crc kubenswrapper[4824]: I1209 10:05:29.422207 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-b846b8d48-dhtnn"] Dec 09 10:05:29 crc kubenswrapper[4824]: W1209 10:05:29.430889 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebb1c8e1_3960_4c38_969b_c483143c87aa.slice/crio-e12cbb135bd5dc48707a9b038d6b5dcfe4717bfb0734533deca21f8101d8217e WatchSource:0}: Error finding container e12cbb135bd5dc48707a9b038d6b5dcfe4717bfb0734533deca21f8101d8217e: Status 404 returned error can't find the container with id e12cbb135bd5dc48707a9b038d6b5dcfe4717bfb0734533deca21f8101d8217e Dec 09 10:05:30 crc kubenswrapper[4824]: I1209 10:05:30.536008 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-b846b8d48-dhtnn" event={"ID":"ebb1c8e1-3960-4c38-969b-c483143c87aa","Type":"ContainerStarted","Data":"e12cbb135bd5dc48707a9b038d6b5dcfe4717bfb0734533deca21f8101d8217e"} Dec 09 10:05:32 crc kubenswrapper[4824]: I1209 10:05:32.343683 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-8rvmd" podUID="bd655e5b-2ac3-4fe8-ad58-6012d0180897" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.96:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 10:05:32 crc kubenswrapper[4824]: I1209 10:05:32.345991 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-8rvmd" podUID="bd655e5b-2ac3-4fe8-ad58-6012d0180897" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.96:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 10:05:39 crc kubenswrapper[4824]: I1209 10:05:39.925630 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-b846b8d48-dhtnn" event={"ID":"ebb1c8e1-3960-4c38-969b-c483143c87aa","Type":"ContainerStarted","Data":"aeb6d2aaccf709b645c37d4ccf2c3e3cde35e6ec8754dbcda60bec1d304afad4"} Dec 09 10:05:39 crc kubenswrapper[4824]: I1209 10:05:39.926273 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-b846b8d48-dhtnn" Dec 09 10:05:39 crc kubenswrapper[4824]: I1209 10:05:39.955657 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-b846b8d48-dhtnn" podStartSLOduration=2.916623874 podStartE2EDuration="12.955638856s" podCreationTimestamp="2025-12-09 10:05:27 +0000 UTC" firstStartedPulling="2025-12-09 10:05:29.43600748 +0000 UTC m=+1085.770512147" lastFinishedPulling="2025-12-09 10:05:39.475022462 +0000 UTC m=+1095.809527129" observedRunningTime="2025-12-09 10:05:39.953520899 +0000 UTC m=+1096.288025556" watchObservedRunningTime="2025-12-09 10:05:39.955638856 +0000 UTC m=+1096.290143523" Dec 09 10:05:48 crc kubenswrapper[4824]: I1209 10:05:48.372435 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-b846b8d48-dhtnn" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.446443 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-6mfg8"] Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.449107 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6mfg8" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.451454 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-xjdln" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.471173 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-lpf4s"] Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.473252 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-lpf4s" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.481937 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-sckbq" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.501965 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-6mfg8"] Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.512990 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2w6b\" (UniqueName: \"kubernetes.io/projected/270a1cf5-f943-4749-8c9a-1c6adf3c45b4-kube-api-access-c2w6b\") pod \"cinder-operator-controller-manager-6c677c69b-lpf4s\" (UID: \"270a1cf5-f943-4749-8c9a-1c6adf3c45b4\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-lpf4s" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.513069 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbzv9\" (UniqueName: \"kubernetes.io/projected/5b34da01-28d8-476c-a2a5-a098489d90c1-kube-api-access-lbzv9\") pod \"barbican-operator-controller-manager-7d9dfd778-6mfg8\" (UID: \"5b34da01-28d8-476c-a2a5-a098489d90c1\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6mfg8" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.519635 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-lpf4s"] Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.543132 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4"] Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.545122 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.551908 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-q77pl" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.554931 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-vx285"] Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.556489 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vx285" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.559400 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-7w9jj" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.574050 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4"] Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.586076 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-vx285"] Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.612837 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-bf7pj"] Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.614587 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-bf7pj" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.615958 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2w6b\" (UniqueName: \"kubernetes.io/projected/270a1cf5-f943-4749-8c9a-1c6adf3c45b4-kube-api-access-c2w6b\") pod \"cinder-operator-controller-manager-6c677c69b-lpf4s\" (UID: \"270a1cf5-f943-4749-8c9a-1c6adf3c45b4\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-lpf4s" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.616013 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph5x7\" (UniqueName: \"kubernetes.io/projected/143e48cd-c956-448c-8dcb-5858a582e29c-kube-api-access-ph5x7\") pod \"designate-operator-controller-manager-697fb699cf-n45n4\" (UID: \"143e48cd-c956-448c-8dcb-5858a582e29c\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.616041 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbzv9\" (UniqueName: \"kubernetes.io/projected/5b34da01-28d8-476c-a2a5-a098489d90c1-kube-api-access-lbzv9\") pod \"barbican-operator-controller-manager-7d9dfd778-6mfg8\" (UID: \"5b34da01-28d8-476c-a2a5-a098489d90c1\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6mfg8" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.616071 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gmrg\" (UniqueName: \"kubernetes.io/projected/54693584-08f2-4326-8062-92d68899d7e4-kube-api-access-5gmrg\") pod \"glance-operator-controller-manager-5697bb5779-vx285\" (UID: \"54693584-08f2-4326-8062-92d68899d7e4\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vx285" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.623766 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-jh9xr" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.624502 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-k92kx"] Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.626392 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-k92kx" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.631898 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-bf7pj"] Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.636244 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-k92kx"] Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.651587 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-hprbh" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.664593 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg"] Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.666373 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.672516 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-t929g" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.683204 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.698247 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbzv9\" (UniqueName: \"kubernetes.io/projected/5b34da01-28d8-476c-a2a5-a098489d90c1-kube-api-access-lbzv9\") pod \"barbican-operator-controller-manager-7d9dfd778-6mfg8\" (UID: \"5b34da01-28d8-476c-a2a5-a098489d90c1\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6mfg8" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.701076 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2w6b\" (UniqueName: \"kubernetes.io/projected/270a1cf5-f943-4749-8c9a-1c6adf3c45b4-kube-api-access-c2w6b\") pod \"cinder-operator-controller-manager-6c677c69b-lpf4s\" (UID: \"270a1cf5-f943-4749-8c9a-1c6adf3c45b4\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-lpf4s" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.702233 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg"] Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.717951 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph5x7\" (UniqueName: \"kubernetes.io/projected/143e48cd-c956-448c-8dcb-5858a582e29c-kube-api-access-ph5x7\") pod \"designate-operator-controller-manager-697fb699cf-n45n4\" (UID: \"143e48cd-c956-448c-8dcb-5858a582e29c\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.718012 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5sx5\" (UniqueName: \"kubernetes.io/projected/30f20f09-f33f-4c7a-a4b4-41b51f73d692-kube-api-access-q5sx5\") pod \"horizon-operator-controller-manager-68c6d99b8f-k92kx\" (UID: \"30f20f09-f33f-4c7a-a4b4-41b51f73d692\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-k92kx" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.718046 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gmrg\" (UniqueName: \"kubernetes.io/projected/54693584-08f2-4326-8062-92d68899d7e4-kube-api-access-5gmrg\") pod \"glance-operator-controller-manager-5697bb5779-vx285\" (UID: \"54693584-08f2-4326-8062-92d68899d7e4\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vx285" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.718099 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97-cert\") pod \"infra-operator-controller-manager-78d48bff9d-p5zwg\" (UID: \"1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.718145 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5b27\" (UniqueName: \"kubernetes.io/projected/1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97-kube-api-access-n5b27\") pod \"infra-operator-controller-manager-78d48bff9d-p5zwg\" (UID: \"1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.718177 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7wtx\" (UniqueName: \"kubernetes.io/projected/ccc192e8-2648-4b1b-8420-1cbd1a27a916-kube-api-access-x7wtx\") pod \"heat-operator-controller-manager-5f64f6f8bb-bf7pj\" (UID: \"ccc192e8-2648-4b1b-8420-1cbd1a27a916\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-bf7pj" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.719215 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-lfprd"] Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.736432 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-lfprd"] Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.736564 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-lfprd" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.739109 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-424j9" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.746622 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-4q2h7"] Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.748417 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4q2h7" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.754839 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-hp7qr" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.757726 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9z2kx"] Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.759683 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9z2kx" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.764642 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-c4lp7" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.774183 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gmrg\" (UniqueName: \"kubernetes.io/projected/54693584-08f2-4326-8062-92d68899d7e4-kube-api-access-5gmrg\") pod \"glance-operator-controller-manager-5697bb5779-vx285\" (UID: \"54693584-08f2-4326-8062-92d68899d7e4\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vx285" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.783400 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph5x7\" (UniqueName: \"kubernetes.io/projected/143e48cd-c956-448c-8dcb-5858a582e29c-kube-api-access-ph5x7\") pod \"designate-operator-controller-manager-697fb699cf-n45n4\" (UID: \"143e48cd-c956-448c-8dcb-5858a582e29c\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.788102 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6mfg8" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.807756 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-lpf4s" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.820167 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-4q2h7"] Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.821308 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxmpl\" (UniqueName: \"kubernetes.io/projected/d1323625-eb0f-4606-bee7-7acbd453d4a5-kube-api-access-rxmpl\") pod \"manila-operator-controller-manager-5b5fd79c9c-9z2kx\" (UID: \"d1323625-eb0f-4606-bee7-7acbd453d4a5\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9z2kx" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.821350 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5sx5\" (UniqueName: \"kubernetes.io/projected/30f20f09-f33f-4c7a-a4b4-41b51f73d692-kube-api-access-q5sx5\") pod \"horizon-operator-controller-manager-68c6d99b8f-k92kx\" (UID: \"30f20f09-f33f-4c7a-a4b4-41b51f73d692\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-k92kx" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.821395 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97-cert\") pod \"infra-operator-controller-manager-78d48bff9d-p5zwg\" (UID: \"1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.821488 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8ql8\" (UniqueName: \"kubernetes.io/projected/7de1c7e2-adcb-4311-926c-ffc42e3e5fde-kube-api-access-f8ql8\") pod \"keystone-operator-controller-manager-7765d96ddf-4q2h7\" (UID: \"7de1c7e2-adcb-4311-926c-ffc42e3e5fde\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4q2h7" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.821510 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5b27\" (UniqueName: \"kubernetes.io/projected/1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97-kube-api-access-n5b27\") pod \"infra-operator-controller-manager-78d48bff9d-p5zwg\" (UID: \"1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.821537 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7wtx\" (UniqueName: \"kubernetes.io/projected/ccc192e8-2648-4b1b-8420-1cbd1a27a916-kube-api-access-x7wtx\") pod \"heat-operator-controller-manager-5f64f6f8bb-bf7pj\" (UID: \"ccc192e8-2648-4b1b-8420-1cbd1a27a916\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-bf7pj" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.821559 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvvpk\" (UniqueName: \"kubernetes.io/projected/b0c20e13-b233-4b21-8f47-facef9603735-kube-api-access-qvvpk\") pod \"ironic-operator-controller-manager-967d97867-lfprd\" (UID: \"b0c20e13-b233-4b21-8f47-facef9603735\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-lfprd" Dec 09 10:06:10 crc kubenswrapper[4824]: E1209 10:06:10.822015 4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 09 10:06:10 crc kubenswrapper[4824]: E1209 10:06:10.822064 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97-cert podName:1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97 nodeName:}" failed. No retries permitted until 2025-12-09 10:06:11.322042039 +0000 UTC m=+1127.656546706 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97-cert") pod "infra-operator-controller-manager-78d48bff9d-p5zwg" (UID: "1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97") : secret "infra-operator-webhook-server-cert" not found Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.871893 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5b27\" (UniqueName: \"kubernetes.io/projected/1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97-kube-api-access-n5b27\") pod \"infra-operator-controller-manager-78d48bff9d-p5zwg\" (UID: \"1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.873902 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl"] Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.874901 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5sx5\" (UniqueName: \"kubernetes.io/projected/30f20f09-f33f-4c7a-a4b4-41b51f73d692-kube-api-access-q5sx5\") pod \"horizon-operator-controller-manager-68c6d99b8f-k92kx\" (UID: \"30f20f09-f33f-4c7a-a4b4-41b51f73d692\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-k92kx" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.876326 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.878491 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.887951 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7wtx\" (UniqueName: \"kubernetes.io/projected/ccc192e8-2648-4b1b-8420-1cbd1a27a916-kube-api-access-x7wtx\") pod \"heat-operator-controller-manager-5f64f6f8bb-bf7pj\" (UID: \"ccc192e8-2648-4b1b-8420-1cbd1a27a916\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-bf7pj" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.892585 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-gspll" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.894557 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vx285" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.919014 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9z2kx"] Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.923464 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxmpl\" (UniqueName: \"kubernetes.io/projected/d1323625-eb0f-4606-bee7-7acbd453d4a5-kube-api-access-rxmpl\") pod \"manila-operator-controller-manager-5b5fd79c9c-9z2kx\" (UID: \"d1323625-eb0f-4606-bee7-7acbd453d4a5\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9z2kx" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.923651 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8ql8\" (UniqueName: \"kubernetes.io/projected/7de1c7e2-adcb-4311-926c-ffc42e3e5fde-kube-api-access-f8ql8\") pod \"keystone-operator-controller-manager-7765d96ddf-4q2h7\" (UID: \"7de1c7e2-adcb-4311-926c-ffc42e3e5fde\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4q2h7" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.923705 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvvpk\" (UniqueName: \"kubernetes.io/projected/b0c20e13-b233-4b21-8f47-facef9603735-kube-api-access-qvvpk\") pod \"ironic-operator-controller-manager-967d97867-lfprd\" (UID: \"b0c20e13-b233-4b21-8f47-facef9603735\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-lfprd" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.923747 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klpl6\" (UniqueName: \"kubernetes.io/projected/b6577050-6a21-43bb-84aa-20aad247aafc-kube-api-access-klpl6\") pod \"mariadb-operator-controller-manager-79c8c4686c-r9fcl\" (UID: \"b6577050-6a21-43bb-84aa-20aad247aafc\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.989141 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxmpl\" (UniqueName: \"kubernetes.io/projected/d1323625-eb0f-4606-bee7-7acbd453d4a5-kube-api-access-rxmpl\") pod \"manila-operator-controller-manager-5b5fd79c9c-9z2kx\" (UID: \"d1323625-eb0f-4606-bee7-7acbd453d4a5\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9z2kx" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.989213 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl"] Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.989523 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-k92kx" Dec 09 10:06:10 crc kubenswrapper[4824]: I1209 10:06:10.991060 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-bf7pj" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.002277 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8ql8\" (UniqueName: \"kubernetes.io/projected/7de1c7e2-adcb-4311-926c-ffc42e3e5fde-kube-api-access-f8ql8\") pod \"keystone-operator-controller-manager-7765d96ddf-4q2h7\" (UID: \"7de1c7e2-adcb-4311-926c-ffc42e3e5fde\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4q2h7" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.028346 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klpl6\" (UniqueName: \"kubernetes.io/projected/b6577050-6a21-43bb-84aa-20aad247aafc-kube-api-access-klpl6\") pod \"mariadb-operator-controller-manager-79c8c4686c-r9fcl\" (UID: \"b6577050-6a21-43bb-84aa-20aad247aafc\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.047468 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9z2kx" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.130267 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8"] Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.131967 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.140803 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-gpwbk" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.159841 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh"] Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.160833 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvvpk\" (UniqueName: \"kubernetes.io/projected/b0c20e13-b233-4b21-8f47-facef9603735-kube-api-access-qvvpk\") pod \"ironic-operator-controller-manager-967d97867-lfprd\" (UID: \"b0c20e13-b233-4b21-8f47-facef9603735\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-lfprd" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.179595 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh"] Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.183949 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.183958 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.185438 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klpl6\" (UniqueName: \"kubernetes.io/projected/b6577050-6a21-43bb-84aa-20aad247aafc-kube-api-access-klpl6\") pod \"mariadb-operator-controller-manager-79c8c4686c-r9fcl\" (UID: \"b6577050-6a21-43bb-84aa-20aad247aafc\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.188982 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-k8fpj" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.189656 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-8pp7h" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.203944 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-lfprd" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.231857 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d577n\" (UniqueName: \"kubernetes.io/projected/23621a98-1d6b-421a-976c-965cecfe6db1-kube-api-access-d577n\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-hlwv8\" (UID: \"23621a98-1d6b-421a-976c-965cecfe6db1\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.248147 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh"] Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.250643 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4q2h7" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.287616 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8"] Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.343061 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97-cert\") pod \"infra-operator-controller-manager-78d48bff9d-p5zwg\" (UID: \"1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.343209 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbmzm\" (UniqueName: \"kubernetes.io/projected/91bf6cd9-a4fe-46e8-b1e7-39e57c1398b3-kube-api-access-xbmzm\") pod \"octavia-operator-controller-manager-998648c74-5vnrh\" (UID: \"91bf6cd9-a4fe-46e8-b1e7-39e57c1398b3\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.343282 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh8s2\" (UniqueName: \"kubernetes.io/projected/d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e-kube-api-access-qh8s2\") pod \"nova-operator-controller-manager-697bc559fc-rbwkh\" (UID: \"d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.343407 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d577n\" (UniqueName: \"kubernetes.io/projected/23621a98-1d6b-421a-976c-965cecfe6db1-kube-api-access-d577n\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-hlwv8\" (UID: \"23621a98-1d6b-421a-976c-965cecfe6db1\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8" Dec 09 10:06:11 crc kubenswrapper[4824]: E1209 10:06:11.346325 4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 09 10:06:11 crc kubenswrapper[4824]: E1209 10:06:11.346406 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97-cert podName:1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97 nodeName:}" failed. No retries permitted until 2025-12-09 10:06:12.346380508 +0000 UTC m=+1128.680885315 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97-cert") pod "infra-operator-controller-manager-78d48bff9d-p5zwg" (UID: "1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97") : secret "infra-operator-webhook-server-cert" not found Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.375127 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh"] Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.375236 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284"] Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.377498 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.389471 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-xbb4r" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.389827 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.416348 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9"] Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.420057 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.432089 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-wpfmq" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.442293 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9"] Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.446481 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxdn2\" (UniqueName: \"kubernetes.io/projected/c2889de3-49b2-4465-8235-37f9e58b42a3-kube-api-access-bxdn2\") pod \"openstack-baremetal-operator-controller-manager-84b575879ftr284\" (UID: \"c2889de3-49b2-4465-8235-37f9e58b42a3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.446610 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2889de3-49b2-4465-8235-37f9e58b42a3-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879ftr284\" (UID: \"c2889de3-49b2-4465-8235-37f9e58b42a3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.446705 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbmzm\" (UniqueName: \"kubernetes.io/projected/91bf6cd9-a4fe-46e8-b1e7-39e57c1398b3-kube-api-access-xbmzm\") pod \"octavia-operator-controller-manager-998648c74-5vnrh\" (UID: \"91bf6cd9-a4fe-46e8-b1e7-39e57c1398b3\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.446751 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qh8s2\" (UniqueName: \"kubernetes.io/projected/d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e-kube-api-access-qh8s2\") pod \"nova-operator-controller-manager-697bc559fc-rbwkh\" (UID: \"d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.460291 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284"] Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.478980 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq"] Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.482164 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.486079 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.491129 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-p9r7r" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.503892 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl"] Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.505655 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.516456 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-bkbbq" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.549935 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq"] Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.551029 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxdn2\" (UniqueName: \"kubernetes.io/projected/c2889de3-49b2-4465-8235-37f9e58b42a3-kube-api-access-bxdn2\") pod \"openstack-baremetal-operator-controller-manager-84b575879ftr284\" (UID: \"c2889de3-49b2-4465-8235-37f9e58b42a3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.551100 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2889de3-49b2-4465-8235-37f9e58b42a3-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879ftr284\" (UID: \"c2889de3-49b2-4465-8235-37f9e58b42a3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.551159 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thjwf\" (UniqueName: \"kubernetes.io/projected/383a67da-0f83-41e2-82d6-809d1b6cdbec-kube-api-access-thjwf\") pod \"placement-operator-controller-manager-78f8948974-pqhrq\" (UID: \"383a67da-0f83-41e2-82d6-809d1b6cdbec\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.551200 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7zdc\" (UniqueName: \"kubernetes.io/projected/31749799-9eb4-403f-a61d-9d50d8bc8367-kube-api-access-m7zdc\") pod \"ovn-operator-controller-manager-b6456fdb6-z68l9\" (UID: \"31749799-9eb4-403f-a61d-9d50d8bc8367\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.551263 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2k8c\" (UniqueName: \"kubernetes.io/projected/6e91c44f-6221-4dfd-9c18-2bbedff02850-kube-api-access-b2k8c\") pod \"swift-operator-controller-manager-9d58d64bc-rz4sl\" (UID: \"6e91c44f-6221-4dfd-9c18-2bbedff02850\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl" Dec 09 10:06:11 crc kubenswrapper[4824]: E1209 10:06:11.551677 4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 10:06:11 crc kubenswrapper[4824]: E1209 10:06:11.551720 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c2889de3-49b2-4465-8235-37f9e58b42a3-cert podName:c2889de3-49b2-4465-8235-37f9e58b42a3 nodeName:}" failed. No retries permitted until 2025-12-09 10:06:12.051702978 +0000 UTC m=+1128.386207645 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c2889de3-49b2-4465-8235-37f9e58b42a3-cert") pod "openstack-baremetal-operator-controller-manager-84b575879ftr284" (UID: "c2889de3-49b2-4465-8235-37f9e58b42a3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.568690 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d577n\" (UniqueName: \"kubernetes.io/projected/23621a98-1d6b-421a-976c-965cecfe6db1-kube-api-access-d577n\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-hlwv8\" (UID: \"23621a98-1d6b-421a-976c-965cecfe6db1\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.572642 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxdn2\" (UniqueName: \"kubernetes.io/projected/c2889de3-49b2-4465-8235-37f9e58b42a3-kube-api-access-bxdn2\") pod \"openstack-baremetal-operator-controller-manager-84b575879ftr284\" (UID: \"c2889de3-49b2-4465-8235-37f9e58b42a3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.584509 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbmzm\" (UniqueName: \"kubernetes.io/projected/91bf6cd9-a4fe-46e8-b1e7-39e57c1398b3-kube-api-access-xbmzm\") pod \"octavia-operator-controller-manager-998648c74-5vnrh\" (UID: \"91bf6cd9-a4fe-46e8-b1e7-39e57c1398b3\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.584583 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl"] Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.586111 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.587050 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh8s2\" (UniqueName: \"kubernetes.io/projected/d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e-kube-api-access-qh8s2\") pod \"nova-operator-controller-manager-697bc559fc-rbwkh\" (UID: \"d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.590311 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-lfghv" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.659709 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thjwf\" (UniqueName: \"kubernetes.io/projected/383a67da-0f83-41e2-82d6-809d1b6cdbec-kube-api-access-thjwf\") pod \"placement-operator-controller-manager-78f8948974-pqhrq\" (UID: \"383a67da-0f83-41e2-82d6-809d1b6cdbec\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.660709 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7zdc\" (UniqueName: \"kubernetes.io/projected/31749799-9eb4-403f-a61d-9d50d8bc8367-kube-api-access-m7zdc\") pod \"ovn-operator-controller-manager-b6456fdb6-z68l9\" (UID: \"31749799-9eb4-403f-a61d-9d50d8bc8367\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.665090 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl2dz\" (UniqueName: \"kubernetes.io/projected/feab57e6-788b-4f23-9e99-aa248786052d-kube-api-access-gl2dz\") pod \"telemetry-operator-controller-manager-577c7bd649-xtnkl\" (UID: \"feab57e6-788b-4f23-9e99-aa248786052d\") " pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.665353 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2k8c\" (UniqueName: \"kubernetes.io/projected/6e91c44f-6221-4dfd-9c18-2bbedff02850-kube-api-access-b2k8c\") pod \"swift-operator-controller-manager-9d58d64bc-rz4sl\" (UID: \"6e91c44f-6221-4dfd-9c18-2bbedff02850\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.670980 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl"] Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.686470 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl"] Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.687132 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-55kwm"] Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.688766 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.695003 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-dr2ts" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.698747 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-55kwm"] Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.712862 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-q7psb"] Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.714634 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-q7psb" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.726911 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-q7psb"] Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.727121 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-dv692" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.729132 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thjwf\" (UniqueName: \"kubernetes.io/projected/383a67da-0f83-41e2-82d6-809d1b6cdbec-kube-api-access-thjwf\") pod \"placement-operator-controller-manager-78f8948974-pqhrq\" (UID: \"383a67da-0f83-41e2-82d6-809d1b6cdbec\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.771033 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl2dz\" (UniqueName: \"kubernetes.io/projected/feab57e6-788b-4f23-9e99-aa248786052d-kube-api-access-gl2dz\") pod \"telemetry-operator-controller-manager-577c7bd649-xtnkl\" (UID: \"feab57e6-788b-4f23-9e99-aa248786052d\") " pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.771107 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwtqd\" (UniqueName: \"kubernetes.io/projected/bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8-kube-api-access-cwtqd\") pod \"test-operator-controller-manager-5854674fcc-55kwm\" (UID: \"bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.771237 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq56g\" (UniqueName: \"kubernetes.io/projected/0af0249c-ed71-4f92-b1a3-cb6f43b92529-kube-api-access-lq56g\") pod \"watcher-operator-controller-manager-667bd8d554-q7psb\" (UID: \"0af0249c-ed71-4f92-b1a3-cb6f43b92529\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-q7psb" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.775740 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2k8c\" (UniqueName: \"kubernetes.io/projected/6e91c44f-6221-4dfd-9c18-2bbedff02850-kube-api-access-b2k8c\") pod \"swift-operator-controller-manager-9d58d64bc-rz4sl\" (UID: \"6e91c44f-6221-4dfd-9c18-2bbedff02850\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.817550 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7zdc\" (UniqueName: \"kubernetes.io/projected/31749799-9eb4-403f-a61d-9d50d8bc8367-kube-api-access-m7zdc\") pod \"ovn-operator-controller-manager-b6456fdb6-z68l9\" (UID: \"31749799-9eb4-403f-a61d-9d50d8bc8367\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.830095 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.847047 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7"] Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.848807 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.851096 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl2dz\" (UniqueName: \"kubernetes.io/projected/feab57e6-788b-4f23-9e99-aa248786052d-kube-api-access-gl2dz\") pod \"telemetry-operator-controller-manager-577c7bd649-xtnkl\" (UID: \"feab57e6-788b-4f23-9e99-aa248786052d\") " pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.858764 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.867601 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.868674 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-wbthm" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.875067 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq56g\" (UniqueName: \"kubernetes.io/projected/0af0249c-ed71-4f92-b1a3-cb6f43b92529-kube-api-access-lq56g\") pod \"watcher-operator-controller-manager-667bd8d554-q7psb\" (UID: \"0af0249c-ed71-4f92-b1a3-cb6f43b92529\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-q7psb" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.875274 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwtqd\" (UniqueName: \"kubernetes.io/projected/bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8-kube-api-access-cwtqd\") pod \"test-operator-controller-manager-5854674fcc-55kwm\" (UID: \"bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.885547 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.886263 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.896299 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.957498 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq56g\" (UniqueName: \"kubernetes.io/projected/0af0249c-ed71-4f92-b1a3-cb6f43b92529-kube-api-access-lq56g\") pod \"watcher-operator-controller-manager-667bd8d554-q7psb\" (UID: \"0af0249c-ed71-4f92-b1a3-cb6f43b92529\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-q7psb" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.963979 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.966817 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.970311 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwtqd\" (UniqueName: \"kubernetes.io/projected/bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8-kube-api-access-cwtqd\") pod \"test-operator-controller-manager-5854674fcc-55kwm\" (UID: \"bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" Dec 09 10:06:11 crc kubenswrapper[4824]: I1209 10:06:11.989203 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7"] Dec 09 10:06:12 crc kubenswrapper[4824]: I1209 10:06:12.033657 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwtpc"] Dec 09 10:06:12 crc kubenswrapper[4824]: I1209 10:06:12.036201 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwtpc" Dec 09 10:06:12 crc kubenswrapper[4824]: I1209 10:06:12.051935 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwtpc"] Dec 09 10:06:12 crc kubenswrapper[4824]: I1209 10:06:12.054752 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd85g\" (UniqueName: \"kubernetes.io/projected/e7e496dc-23c4-47cd-9c5a-aa4430130849-kube-api-access-cd85g\") pod \"openstack-operator-controller-manager-598df84fd8-l2bt7\" (UID: \"e7e496dc-23c4-47cd-9c5a-aa4430130849\") " pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 10:06:12 crc kubenswrapper[4824]: I1209 10:06:12.054929 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-metrics-certs\") pod \"openstack-operator-controller-manager-598df84fd8-l2bt7\" (UID: \"e7e496dc-23c4-47cd-9c5a-aa4430130849\") " pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 10:06:12 crc kubenswrapper[4824]: I1209 10:06:12.054954 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-6c6k8" Dec 09 10:06:12 crc kubenswrapper[4824]: I1209 10:06:12.055048 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-webhook-certs\") pod \"openstack-operator-controller-manager-598df84fd8-l2bt7\" (UID: \"e7e496dc-23c4-47cd-9c5a-aa4430130849\") " pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 10:06:12 crc kubenswrapper[4824]: I1209 10:06:12.055136 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2889de3-49b2-4465-8235-37f9e58b42a3-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879ftr284\" (UID: \"c2889de3-49b2-4465-8235-37f9e58b42a3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" Dec 09 10:06:12 crc kubenswrapper[4824]: E1209 10:06:12.056396 4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 10:06:12 crc kubenswrapper[4824]: E1209 10:06:12.056469 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c2889de3-49b2-4465-8235-37f9e58b42a3-cert podName:c2889de3-49b2-4465-8235-37f9e58b42a3 nodeName:}" failed. No retries permitted until 2025-12-09 10:06:13.056441479 +0000 UTC m=+1129.390946146 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c2889de3-49b2-4465-8235-37f9e58b42a3-cert") pod "openstack-baremetal-operator-controller-manager-84b575879ftr284" (UID: "c2889de3-49b2-4465-8235-37f9e58b42a3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 10:06:12 crc kubenswrapper[4824]: I1209 10:06:12.064109 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" Dec 09 10:06:12 crc kubenswrapper[4824]: I1209 10:06:12.087839 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" Dec 09 10:06:12 crc kubenswrapper[4824]: I1209 10:06:12.106414 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-q7psb" Dec 09 10:06:12 crc kubenswrapper[4824]: I1209 10:06:12.157920 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd85g\" (UniqueName: \"kubernetes.io/projected/e7e496dc-23c4-47cd-9c5a-aa4430130849-kube-api-access-cd85g\") pod \"openstack-operator-controller-manager-598df84fd8-l2bt7\" (UID: \"e7e496dc-23c4-47cd-9c5a-aa4430130849\") " pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 10:06:12 crc kubenswrapper[4824]: I1209 10:06:12.158304 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-metrics-certs\") pod \"openstack-operator-controller-manager-598df84fd8-l2bt7\" (UID: \"e7e496dc-23c4-47cd-9c5a-aa4430130849\") " pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 10:06:12 crc kubenswrapper[4824]: I1209 10:06:12.158345 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-webhook-certs\") pod \"openstack-operator-controller-manager-598df84fd8-l2bt7\" (UID: \"e7e496dc-23c4-47cd-9c5a-aa4430130849\") " pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 10:06:12 crc kubenswrapper[4824]: I1209 10:06:12.158458 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmfjn\" (UniqueName: \"kubernetes.io/projected/09d4392b-616f-4531-837e-46f1ef396caf-kube-api-access-lmfjn\") pod \"rabbitmq-cluster-operator-manager-668c99d594-lwtpc\" (UID: \"09d4392b-616f-4531-837e-46f1ef396caf\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwtpc" Dec 09 10:06:12 crc kubenswrapper[4824]: E1209 10:06:12.159587 4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 09 10:06:12 crc kubenswrapper[4824]: E1209 10:06:12.160619 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-metrics-certs podName:e7e496dc-23c4-47cd-9c5a-aa4430130849 nodeName:}" failed. No retries permitted until 2025-12-09 10:06:12.659621556 +0000 UTC m=+1128.994126223 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-metrics-certs") pod "openstack-operator-controller-manager-598df84fd8-l2bt7" (UID: "e7e496dc-23c4-47cd-9c5a-aa4430130849") : secret "metrics-server-cert" not found Dec 09 10:06:12 crc kubenswrapper[4824]: E1209 10:06:12.161009 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 09 10:06:12 crc kubenswrapper[4824]: E1209 10:06:12.161049 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-webhook-certs podName:e7e496dc-23c4-47cd-9c5a-aa4430130849 nodeName:}" failed. No retries permitted until 2025-12-09 10:06:12.661034741 +0000 UTC m=+1128.995539408 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-webhook-certs") pod "openstack-operator-controller-manager-598df84fd8-l2bt7" (UID: "e7e496dc-23c4-47cd-9c5a-aa4430130849") : secret "webhook-server-cert" not found Dec 09 10:06:12 crc kubenswrapper[4824]: I1209 10:06:12.231959 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-6mfg8"] Dec 09 10:06:12 crc kubenswrapper[4824]: I1209 10:06:12.262651 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmfjn\" (UniqueName: \"kubernetes.io/projected/09d4392b-616f-4531-837e-46f1ef396caf-kube-api-access-lmfjn\") pod \"rabbitmq-cluster-operator-manager-668c99d594-lwtpc\" (UID: \"09d4392b-616f-4531-837e-46f1ef396caf\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwtpc" Dec 09 10:06:12 crc kubenswrapper[4824]: I1209 10:06:12.323738 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd85g\" (UniqueName: \"kubernetes.io/projected/e7e496dc-23c4-47cd-9c5a-aa4430130849-kube-api-access-cd85g\") pod \"openstack-operator-controller-manager-598df84fd8-l2bt7\" (UID: \"e7e496dc-23c4-47cd-9c5a-aa4430130849\") " pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 10:06:12 crc kubenswrapper[4824]: I1209 10:06:12.324523 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmfjn\" (UniqueName: \"kubernetes.io/projected/09d4392b-616f-4531-837e-46f1ef396caf-kube-api-access-lmfjn\") pod \"rabbitmq-cluster-operator-manager-668c99d594-lwtpc\" (UID: \"09d4392b-616f-4531-837e-46f1ef396caf\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwtpc" Dec 09 10:06:12 crc kubenswrapper[4824]: I1209 10:06:12.327535 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwtpc" Dec 09 10:06:12 crc kubenswrapper[4824]: I1209 10:06:12.374948 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97-cert\") pod \"infra-operator-controller-manager-78d48bff9d-p5zwg\" (UID: \"1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" Dec 09 10:06:12 crc kubenswrapper[4824]: E1209 10:06:12.375249 4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 09 10:06:12 crc kubenswrapper[4824]: E1209 10:06:12.375370 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97-cert podName:1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97 nodeName:}" failed. No retries permitted until 2025-12-09 10:06:14.375341055 +0000 UTC m=+1130.709845722 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97-cert") pod "infra-operator-controller-manager-78d48bff9d-p5zwg" (UID: "1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97") : secret "infra-operator-webhook-server-cert" not found Dec 09 10:06:12 crc kubenswrapper[4824]: I1209 10:06:12.735354 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-metrics-certs\") pod \"openstack-operator-controller-manager-598df84fd8-l2bt7\" (UID: \"e7e496dc-23c4-47cd-9c5a-aa4430130849\") " pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 10:06:12 crc kubenswrapper[4824]: I1209 10:06:12.735442 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-webhook-certs\") pod \"openstack-operator-controller-manager-598df84fd8-l2bt7\" (UID: \"e7e496dc-23c4-47cd-9c5a-aa4430130849\") " pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 10:06:12 crc kubenswrapper[4824]: E1209 10:06:12.735653 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 09 10:06:12 crc kubenswrapper[4824]: E1209 10:06:12.735722 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-webhook-certs podName:e7e496dc-23c4-47cd-9c5a-aa4430130849 nodeName:}" failed. No retries permitted until 2025-12-09 10:06:13.735698558 +0000 UTC m=+1130.070203225 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-webhook-certs") pod "openstack-operator-controller-manager-598df84fd8-l2bt7" (UID: "e7e496dc-23c4-47cd-9c5a-aa4430130849") : secret "webhook-server-cert" not found Dec 09 10:06:12 crc kubenswrapper[4824]: E1209 10:06:12.735740 4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 09 10:06:12 crc kubenswrapper[4824]: E1209 10:06:12.735846 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-metrics-certs podName:e7e496dc-23c4-47cd-9c5a-aa4430130849 nodeName:}" failed. No retries permitted until 2025-12-09 10:06:13.735820302 +0000 UTC m=+1130.070325149 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-metrics-certs") pod "openstack-operator-controller-manager-598df84fd8-l2bt7" (UID: "e7e496dc-23c4-47cd-9c5a-aa4430130849") : secret "metrics-server-cert" not found Dec 09 10:06:12 crc kubenswrapper[4824]: I1209 10:06:12.756931 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6mfg8" event={"ID":"5b34da01-28d8-476c-a2a5-a098489d90c1","Type":"ContainerStarted","Data":"96c38cdc4dacf9eb78ebadb07b048ec35d43af3a526b0c50369a56746468d0f3"} Dec 09 10:06:13 crc kubenswrapper[4824]: I1209 10:06:13.138708 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2889de3-49b2-4465-8235-37f9e58b42a3-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879ftr284\" (UID: \"c2889de3-49b2-4465-8235-37f9e58b42a3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" Dec 09 10:06:13 crc kubenswrapper[4824]: E1209 10:06:13.141560 4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 10:06:13 crc kubenswrapper[4824]: E1209 10:06:13.141627 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c2889de3-49b2-4465-8235-37f9e58b42a3-cert podName:c2889de3-49b2-4465-8235-37f9e58b42a3 nodeName:}" failed. No retries permitted until 2025-12-09 10:06:15.141604929 +0000 UTC m=+1131.476109596 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c2889de3-49b2-4465-8235-37f9e58b42a3-cert") pod "openstack-baremetal-operator-controller-manager-84b575879ftr284" (UID: "c2889de3-49b2-4465-8235-37f9e58b42a3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 10:06:13 crc kubenswrapper[4824]: I1209 10:06:13.745246 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-metrics-certs\") pod \"openstack-operator-controller-manager-598df84fd8-l2bt7\" (UID: \"e7e496dc-23c4-47cd-9c5a-aa4430130849\") " pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 10:06:13 crc kubenswrapper[4824]: I1209 10:06:13.746343 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-webhook-certs\") pod \"openstack-operator-controller-manager-598df84fd8-l2bt7\" (UID: \"e7e496dc-23c4-47cd-9c5a-aa4430130849\") " pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 10:06:13 crc kubenswrapper[4824]: E1209 10:06:13.745410 4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 09 10:06:13 crc kubenswrapper[4824]: E1209 10:06:13.746631 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-metrics-certs podName:e7e496dc-23c4-47cd-9c5a-aa4430130849 nodeName:}" failed. No retries permitted until 2025-12-09 10:06:15.746606795 +0000 UTC m=+1132.081111462 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-metrics-certs") pod "openstack-operator-controller-manager-598df84fd8-l2bt7" (UID: "e7e496dc-23c4-47cd-9c5a-aa4430130849") : secret "metrics-server-cert" not found Dec 09 10:06:13 crc kubenswrapper[4824]: E1209 10:06:13.746471 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 09 10:06:13 crc kubenswrapper[4824]: E1209 10:06:13.746800 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-webhook-certs podName:e7e496dc-23c4-47cd-9c5a-aa4430130849 nodeName:}" failed. No retries permitted until 2025-12-09 10:06:15.746748439 +0000 UTC m=+1132.081253266 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-webhook-certs") pod "openstack-operator-controller-manager-598df84fd8-l2bt7" (UID: "e7e496dc-23c4-47cd-9c5a-aa4430130849") : secret "webhook-server-cert" not found Dec 09 10:06:14 crc kubenswrapper[4824]: I1209 10:06:14.389190 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97-cert\") pod \"infra-operator-controller-manager-78d48bff9d-p5zwg\" (UID: \"1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" Dec 09 10:06:14 crc kubenswrapper[4824]: E1209 10:06:14.389814 4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 09 10:06:14 crc kubenswrapper[4824]: E1209 10:06:14.389879 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97-cert podName:1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97 nodeName:}" failed. No retries permitted until 2025-12-09 10:06:18.389857787 +0000 UTC m=+1134.724362444 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97-cert") pod "infra-operator-controller-manager-78d48bff9d-p5zwg" (UID: "1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97") : secret "infra-operator-webhook-server-cert" not found Dec 09 10:06:15 crc kubenswrapper[4824]: I1209 10:06:15.210783 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2889de3-49b2-4465-8235-37f9e58b42a3-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879ftr284\" (UID: \"c2889de3-49b2-4465-8235-37f9e58b42a3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" Dec 09 10:06:15 crc kubenswrapper[4824]: E1209 10:06:15.211357 4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 10:06:15 crc kubenswrapper[4824]: E1209 10:06:15.211439 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c2889de3-49b2-4465-8235-37f9e58b42a3-cert podName:c2889de3-49b2-4465-8235-37f9e58b42a3 nodeName:}" failed. No retries permitted until 2025-12-09 10:06:19.211408787 +0000 UTC m=+1135.545913454 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c2889de3-49b2-4465-8235-37f9e58b42a3-cert") pod "openstack-baremetal-operator-controller-manager-84b575879ftr284" (UID: "c2889de3-49b2-4465-8235-37f9e58b42a3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 10:06:15 crc kubenswrapper[4824]: I1209 10:06:15.219455 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-lpf4s"] Dec 09 10:06:15 crc kubenswrapper[4824]: W1209 10:06:15.256435 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod270a1cf5_f943_4749_8c9a_1c6adf3c45b4.slice/crio-0c44566de4ed27104dbf2492b627236e82aa032ad7cb6242a2957d70c765ebbe WatchSource:0}: Error finding container 0c44566de4ed27104dbf2492b627236e82aa032ad7cb6242a2957d70c765ebbe: Status 404 returned error can't find the container with id 0c44566de4ed27104dbf2492b627236e82aa032ad7cb6242a2957d70c765ebbe Dec 09 10:06:15 crc kubenswrapper[4824]: I1209 10:06:15.263914 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4"] Dec 09 10:06:15 crc kubenswrapper[4824]: I1209 10:06:15.276131 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-bf7pj"] Dec 09 10:06:15 crc kubenswrapper[4824]: I1209 10:06:15.286998 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-4q2h7"] Dec 09 10:06:15 crc kubenswrapper[4824]: I1209 10:06:15.304410 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-vx285"] Dec 09 10:06:15 crc kubenswrapper[4824]: I1209 10:06:15.317425 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-k92kx"] Dec 09 10:06:15 crc kubenswrapper[4824]: W1209 10:06:15.319882 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7de1c7e2_adcb_4311_926c_ffc42e3e5fde.slice/crio-30720b80ad2ff71aa1bd68eaccd7369eed93b8f747aa66b7d5a770fd378fc686 WatchSource:0}: Error finding container 30720b80ad2ff71aa1bd68eaccd7369eed93b8f747aa66b7d5a770fd378fc686: Status 404 returned error can't find the container with id 30720b80ad2ff71aa1bd68eaccd7369eed93b8f747aa66b7d5a770fd378fc686 Dec 09 10:06:15 crc kubenswrapper[4824]: I1209 10:06:15.331765 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9z2kx"] Dec 09 10:06:15 crc kubenswrapper[4824]: I1209 10:06:15.421585 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl"] Dec 09 10:06:15 crc kubenswrapper[4824]: I1209 10:06:15.452642 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-lfprd"] Dec 09 10:06:15 crc kubenswrapper[4824]: I1209 10:06:15.800155 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl" event={"ID":"b6577050-6a21-43bb-84aa-20aad247aafc","Type":"ContainerStarted","Data":"53ba18b958edd519d9d36d2c270893d6fc30b80c2fdec2bd91b44cb7bcc65fa8"} Dec 09 10:06:15 crc kubenswrapper[4824]: I1209 10:06:15.828022 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-bf7pj" event={"ID":"ccc192e8-2648-4b1b-8420-1cbd1a27a916","Type":"ContainerStarted","Data":"a136365048ad885d313d9fe5e333d326334e987fe95be159e7d5619287c246fa"} Dec 09 10:06:15 crc kubenswrapper[4824]: I1209 10:06:15.844025 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-metrics-certs\") pod \"openstack-operator-controller-manager-598df84fd8-l2bt7\" (UID: \"e7e496dc-23c4-47cd-9c5a-aa4430130849\") " pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 10:06:15 crc kubenswrapper[4824]: E1209 10:06:15.844196 4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 09 10:06:15 crc kubenswrapper[4824]: E1209 10:06:15.844560 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-metrics-certs podName:e7e496dc-23c4-47cd-9c5a-aa4430130849 nodeName:}" failed. No retries permitted until 2025-12-09 10:06:19.84451266 +0000 UTC m=+1136.179017327 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-metrics-certs") pod "openstack-operator-controller-manager-598df84fd8-l2bt7" (UID: "e7e496dc-23c4-47cd-9c5a-aa4430130849") : secret "metrics-server-cert" not found Dec 09 10:06:15 crc kubenswrapper[4824]: E1209 10:06:15.844665 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 09 10:06:15 crc kubenswrapper[4824]: E1209 10:06:15.844721 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-webhook-certs podName:e7e496dc-23c4-47cd-9c5a-aa4430130849 nodeName:}" failed. No retries permitted until 2025-12-09 10:06:19.844712956 +0000 UTC m=+1136.179217623 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-webhook-certs") pod "openstack-operator-controller-manager-598df84fd8-l2bt7" (UID: "e7e496dc-23c4-47cd-9c5a-aa4430130849") : secret "webhook-server-cert" not found Dec 09 10:06:15 crc kubenswrapper[4824]: I1209 10:06:15.844445 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-webhook-certs\") pod \"openstack-operator-controller-manager-598df84fd8-l2bt7\" (UID: \"e7e496dc-23c4-47cd-9c5a-aa4430130849\") " pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 10:06:15 crc kubenswrapper[4824]: I1209 10:06:15.868217 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-lfprd" event={"ID":"b0c20e13-b233-4b21-8f47-facef9603735","Type":"ContainerStarted","Data":"844bd3b20f9cb322092ac7905735e94d6dfd126ceccf5094f17cb527bf131807"} Dec 09 10:06:15 crc kubenswrapper[4824]: I1209 10:06:15.890081 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-lpf4s" event={"ID":"270a1cf5-f943-4749-8c9a-1c6adf3c45b4","Type":"ContainerStarted","Data":"0c44566de4ed27104dbf2492b627236e82aa032ad7cb6242a2957d70c765ebbe"} Dec 09 10:06:15 crc kubenswrapper[4824]: I1209 10:06:15.892444 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4q2h7" event={"ID":"7de1c7e2-adcb-4311-926c-ffc42e3e5fde","Type":"ContainerStarted","Data":"30720b80ad2ff71aa1bd68eaccd7369eed93b8f747aa66b7d5a770fd378fc686"} Dec 09 10:06:16 crc kubenswrapper[4824]: I1209 10:06:16.049336 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-q7psb"] Dec 09 10:06:16 crc kubenswrapper[4824]: I1209 10:06:16.049378 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh"] Dec 09 10:06:16 crc kubenswrapper[4824]: I1209 10:06:16.049395 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-k92kx" event={"ID":"30f20f09-f33f-4c7a-a4b4-41b51f73d692","Type":"ContainerStarted","Data":"4c52a36a4ee9fd9e8699fd6c9a8aa51d1928a486059d455b8fb533caf6a5004d"} Dec 09 10:06:16 crc kubenswrapper[4824]: I1209 10:06:16.049418 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8"] Dec 09 10:06:16 crc kubenswrapper[4824]: I1209 10:06:16.049439 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9z2kx" event={"ID":"d1323625-eb0f-4606-bee7-7acbd453d4a5","Type":"ContainerStarted","Data":"37db73e2fefe472dbb40640834683a196e6004fd780bc577bbaa424c507fbe6f"} Dec 09 10:06:16 crc kubenswrapper[4824]: I1209 10:06:16.049453 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4" event={"ID":"143e48cd-c956-448c-8dcb-5858a582e29c","Type":"ContainerStarted","Data":"8472e68d56b6f07d4edababc7bfefc50c49b388b60f311444f36a7c7dcd88c8c"} Dec 09 10:06:16 crc kubenswrapper[4824]: I1209 10:06:16.049467 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vx285" event={"ID":"54693584-08f2-4326-8062-92d68899d7e4","Type":"ContainerStarted","Data":"f0dfd80a10a1a01c491083090819085a3a2dc514a9baf1075ac0f80b328507f2"} Dec 09 10:06:16 crc kubenswrapper[4824]: I1209 10:06:16.049483 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl"] Dec 09 10:06:16 crc kubenswrapper[4824]: I1209 10:06:16.057007 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh"] Dec 09 10:06:16 crc kubenswrapper[4824]: I1209 10:06:16.146110 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq"] Dec 09 10:06:16 crc kubenswrapper[4824]: E1209 10:06:16.147426 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-thjwf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-pqhrq_openstack-operators(383a67da-0f83-41e2-82d6-809d1b6cdbec): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 10:06:16 crc kubenswrapper[4824]: E1209 10:06:16.149634 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-thjwf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-pqhrq_openstack-operators(383a67da-0f83-41e2-82d6-809d1b6cdbec): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 10:06:16 crc kubenswrapper[4824]: E1209 10:06:16.151120 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" podUID="383a67da-0f83-41e2-82d6-809d1b6cdbec" Dec 09 10:06:16 crc kubenswrapper[4824]: I1209 10:06:16.161655 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-55kwm"] Dec 09 10:06:16 crc kubenswrapper[4824]: W1209 10:06:16.194414 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfeab57e6_788b_4f23_9e99_aa248786052d.slice/crio-c2e983cd1dba67e15b56af8ad77d6d4103fc84d85a52f638c2f0b7fcfddc9d29 WatchSource:0}: Error finding container c2e983cd1dba67e15b56af8ad77d6d4103fc84d85a52f638c2f0b7fcfddc9d29: Status 404 returned error can't find the container with id c2e983cd1dba67e15b56af8ad77d6d4103fc84d85a52f638c2f0b7fcfddc9d29 Dec 09 10:06:16 crc kubenswrapper[4824]: I1209 10:06:16.197144 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl"] Dec 09 10:06:16 crc kubenswrapper[4824]: E1209 10:06:16.199830 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.20:5001/openstack-k8s-operators/telemetry-operator:adc44e0841ee1c9c853afa02c8f62dffad25da1d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gl2dz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-577c7bd649-xtnkl_openstack-operators(feab57e6-788b-4f23-9e99-aa248786052d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 10:06:16 crc kubenswrapper[4824]: E1209 10:06:16.202289 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gl2dz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-577c7bd649-xtnkl_openstack-operators(feab57e6-788b-4f23-9e99-aa248786052d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 10:06:16 crc kubenswrapper[4824]: E1209 10:06:16.204676 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" podUID="feab57e6-788b-4f23-9e99-aa248786052d" Dec 09 10:06:16 crc kubenswrapper[4824]: E1209 10:06:16.215479 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cwtqd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-55kwm_openstack-operators(bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 10:06:16 crc kubenswrapper[4824]: E1209 10:06:16.217318 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m7zdc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-z68l9_openstack-operators(31749799-9eb4-403f-a61d-9d50d8bc8367): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 10:06:16 crc kubenswrapper[4824]: E1209 10:06:16.218213 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cwtqd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-55kwm_openstack-operators(bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 10:06:16 crc kubenswrapper[4824]: E1209 10:06:16.219317 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" podUID="bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8" Dec 09 10:06:16 crc kubenswrapper[4824]: E1209 10:06:16.219519 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m7zdc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-z68l9_openstack-operators(31749799-9eb4-403f-a61d-9d50d8bc8367): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 10:06:16 crc kubenswrapper[4824]: E1209 10:06:16.220845 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9" podUID="31749799-9eb4-403f-a61d-9d50d8bc8367" Dec 09 10:06:16 crc kubenswrapper[4824]: I1209 10:06:16.234259 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9"] Dec 09 10:06:16 crc kubenswrapper[4824]: E1209 10:06:16.237133 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lmfjn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-lwtpc_openstack-operators(09d4392b-616f-4531-837e-46f1ef396caf): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 10:06:16 crc kubenswrapper[4824]: E1209 10:06:16.238491 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwtpc" podUID="09d4392b-616f-4531-837e-46f1ef396caf" Dec 09 10:06:16 crc kubenswrapper[4824]: I1209 10:06:16.250480 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwtpc"] Dec 09 10:06:17 crc kubenswrapper[4824]: I1209 10:06:17.079203 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" event={"ID":"feab57e6-788b-4f23-9e99-aa248786052d","Type":"ContainerStarted","Data":"c2e983cd1dba67e15b56af8ad77d6d4103fc84d85a52f638c2f0b7fcfddc9d29"} Dec 09 10:06:17 crc kubenswrapper[4824]: E1209 10:06:17.142108 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.20:5001/openstack-k8s-operators/telemetry-operator:adc44e0841ee1c9c853afa02c8f62dffad25da1d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" podUID="feab57e6-788b-4f23-9e99-aa248786052d" Dec 09 10:06:17 crc kubenswrapper[4824]: I1209 10:06:17.146877 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl" event={"ID":"6e91c44f-6221-4dfd-9c18-2bbedff02850","Type":"ContainerStarted","Data":"cb77ca973db8b373fba3ea723113d15ba95005a95bf8eab3dac057a3b30c1939"} Dec 09 10:06:17 crc kubenswrapper[4824]: I1209 10:06:17.188699 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9" event={"ID":"31749799-9eb4-403f-a61d-9d50d8bc8367","Type":"ContainerStarted","Data":"257814a9ca4cd021f1d144c013951dc1cec2629dc12876028a601dfc9924be76"} Dec 09 10:06:17 crc kubenswrapper[4824]: E1209 10:06:17.215514 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9" podUID="31749799-9eb4-403f-a61d-9d50d8bc8367" Dec 09 10:06:17 crc kubenswrapper[4824]: I1209 10:06:17.235271 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8" event={"ID":"23621a98-1d6b-421a-976c-965cecfe6db1","Type":"ContainerStarted","Data":"ea4fa9aef2ccbe1b502e41978591bc3865e91ba3e38e999fba01864f66990c90"} Dec 09 10:06:17 crc kubenswrapper[4824]: I1209 10:06:17.277015 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" event={"ID":"bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8","Type":"ContainerStarted","Data":"da33537b272005578ec44a9b14509b2a6db01db9e7c74df42b648f50fd88083d"} Dec 09 10:06:17 crc kubenswrapper[4824]: E1209 10:06:17.303182 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" podUID="bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8" Dec 09 10:06:17 crc kubenswrapper[4824]: I1209 10:06:17.307742 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" event={"ID":"383a67da-0f83-41e2-82d6-809d1b6cdbec","Type":"ContainerStarted","Data":"84e46b19e657339dc52fa1e18fe5129409282521a638ca99881073f3da45651a"} Dec 09 10:06:17 crc kubenswrapper[4824]: E1209 10:06:17.334138 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" podUID="383a67da-0f83-41e2-82d6-809d1b6cdbec" Dec 09 10:06:17 crc kubenswrapper[4824]: I1209 10:06:17.338070 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" event={"ID":"d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e","Type":"ContainerStarted","Data":"52e9becc6477cd77c1fbe39fea11540c9ca7262197117f997824d599eef5e556"} Dec 09 10:06:17 crc kubenswrapper[4824]: I1209 10:06:17.365174 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwtpc" event={"ID":"09d4392b-616f-4531-837e-46f1ef396caf","Type":"ContainerStarted","Data":"93ba7538fd680c339b8074971f166fec647e0bb125f4d3c642685643bc8ee12b"} Dec 09 10:06:17 crc kubenswrapper[4824]: E1209 10:06:17.369967 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwtpc" podUID="09d4392b-616f-4531-837e-46f1ef396caf" Dec 09 10:06:17 crc kubenswrapper[4824]: I1209 10:06:17.383287 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-q7psb" event={"ID":"0af0249c-ed71-4f92-b1a3-cb6f43b92529","Type":"ContainerStarted","Data":"b57bdfb7c0aaf942e49c5ddf7c91c2ac477ee6047a9d3f55cbb405477db98693"} Dec 09 10:06:17 crc kubenswrapper[4824]: I1209 10:06:17.387010 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh" event={"ID":"91bf6cd9-a4fe-46e8-b1e7-39e57c1398b3","Type":"ContainerStarted","Data":"5b44fcec7ad916bc3cdb6cd9c46d3956400a87ca03b895a8f68fa7df9ffc5800"} Dec 09 10:06:18 crc kubenswrapper[4824]: E1209 10:06:18.454805 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwtpc" podUID="09d4392b-616f-4531-837e-46f1ef396caf" Dec 09 10:06:18 crc kubenswrapper[4824]: E1209 10:06:18.455168 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.20:5001/openstack-k8s-operators/telemetry-operator:adc44e0841ee1c9c853afa02c8f62dffad25da1d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" podUID="feab57e6-788b-4f23-9e99-aa248786052d" Dec 09 10:06:18 crc kubenswrapper[4824]: E1209 10:06:18.455314 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9" podUID="31749799-9eb4-403f-a61d-9d50d8bc8367" Dec 09 10:06:18 crc kubenswrapper[4824]: E1209 10:06:18.463769 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" podUID="bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8" Dec 09 10:06:18 crc kubenswrapper[4824]: E1209 10:06:18.463848 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" podUID="383a67da-0f83-41e2-82d6-809d1b6cdbec" Dec 09 10:06:18 crc kubenswrapper[4824]: I1209 10:06:18.486433 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97-cert\") pod \"infra-operator-controller-manager-78d48bff9d-p5zwg\" (UID: \"1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" Dec 09 10:06:18 crc kubenswrapper[4824]: E1209 10:06:18.487041 4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 09 10:06:18 crc kubenswrapper[4824]: E1209 10:06:18.489283 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97-cert podName:1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97 nodeName:}" failed. No retries permitted until 2025-12-09 10:06:26.489249413 +0000 UTC m=+1142.823754250 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97-cert") pod "infra-operator-controller-manager-78d48bff9d-p5zwg" (UID: "1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97") : secret "infra-operator-webhook-server-cert" not found Dec 09 10:06:19 crc kubenswrapper[4824]: I1209 10:06:19.312262 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2889de3-49b2-4465-8235-37f9e58b42a3-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879ftr284\" (UID: \"c2889de3-49b2-4465-8235-37f9e58b42a3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" Dec 09 10:06:19 crc kubenswrapper[4824]: E1209 10:06:19.313112 4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 10:06:19 crc kubenswrapper[4824]: E1209 10:06:19.313350 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c2889de3-49b2-4465-8235-37f9e58b42a3-cert podName:c2889de3-49b2-4465-8235-37f9e58b42a3 nodeName:}" failed. No retries permitted until 2025-12-09 10:06:27.313307463 +0000 UTC m=+1143.647812150 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c2889de3-49b2-4465-8235-37f9e58b42a3-cert") pod "openstack-baremetal-operator-controller-manager-84b575879ftr284" (UID: "c2889de3-49b2-4465-8235-37f9e58b42a3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 10:06:19 crc kubenswrapper[4824]: I1209 10:06:19.932947 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-metrics-certs\") pod \"openstack-operator-controller-manager-598df84fd8-l2bt7\" (UID: \"e7e496dc-23c4-47cd-9c5a-aa4430130849\") " pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 10:06:19 crc kubenswrapper[4824]: I1209 10:06:19.933023 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-webhook-certs\") pod \"openstack-operator-controller-manager-598df84fd8-l2bt7\" (UID: \"e7e496dc-23c4-47cd-9c5a-aa4430130849\") " pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 10:06:19 crc kubenswrapper[4824]: E1209 10:06:19.934560 4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 09 10:06:19 crc kubenswrapper[4824]: E1209 10:06:19.934620 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-metrics-certs podName:e7e496dc-23c4-47cd-9c5a-aa4430130849 nodeName:}" failed. No retries permitted until 2025-12-09 10:06:27.934600002 +0000 UTC m=+1144.269104659 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-metrics-certs") pod "openstack-operator-controller-manager-598df84fd8-l2bt7" (UID: "e7e496dc-23c4-47cd-9c5a-aa4430130849") : secret "metrics-server-cert" not found Dec 09 10:06:19 crc kubenswrapper[4824]: E1209 10:06:19.935038 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 09 10:06:19 crc kubenswrapper[4824]: E1209 10:06:19.935067 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-webhook-certs podName:e7e496dc-23c4-47cd-9c5a-aa4430130849 nodeName:}" failed. No retries permitted until 2025-12-09 10:06:27.935057377 +0000 UTC m=+1144.269562044 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-webhook-certs") pod "openstack-operator-controller-manager-598df84fd8-l2bt7" (UID: "e7e496dc-23c4-47cd-9c5a-aa4430130849") : secret "webhook-server-cert" not found Dec 09 10:06:26 crc kubenswrapper[4824]: I1209 10:06:26.758306 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97-cert\") pod \"infra-operator-controller-manager-78d48bff9d-p5zwg\" (UID: \"1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" Dec 09 10:06:26 crc kubenswrapper[4824]: I1209 10:06:26.815885 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97-cert\") pod \"infra-operator-controller-manager-78d48bff9d-p5zwg\" (UID: \"1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" Dec 09 10:06:26 crc kubenswrapper[4824]: I1209 10:06:26.966139 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-t929g" Dec 09 10:06:26 crc kubenswrapper[4824]: I1209 10:06:26.978654 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" Dec 09 10:06:27 crc kubenswrapper[4824]: I1209 10:06:27.368269 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2889de3-49b2-4465-8235-37f9e58b42a3-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879ftr284\" (UID: \"c2889de3-49b2-4465-8235-37f9e58b42a3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" Dec 09 10:06:27 crc kubenswrapper[4824]: I1209 10:06:27.375175 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2889de3-49b2-4465-8235-37f9e58b42a3-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879ftr284\" (UID: \"c2889de3-49b2-4465-8235-37f9e58b42a3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" Dec 09 10:06:27 crc kubenswrapper[4824]: I1209 10:06:27.525709 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-xbb4r" Dec 09 10:06:27 crc kubenswrapper[4824]: I1209 10:06:27.534336 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" Dec 09 10:06:28 crc kubenswrapper[4824]: I1209 10:06:28.533808 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-metrics-certs\") pod \"openstack-operator-controller-manager-598df84fd8-l2bt7\" (UID: \"e7e496dc-23c4-47cd-9c5a-aa4430130849\") " pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 10:06:28 crc kubenswrapper[4824]: I1209 10:06:28.533933 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-webhook-certs\") pod \"openstack-operator-controller-manager-598df84fd8-l2bt7\" (UID: \"e7e496dc-23c4-47cd-9c5a-aa4430130849\") " pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 10:06:28 crc kubenswrapper[4824]: E1209 10:06:28.534419 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 09 10:06:28 crc kubenswrapper[4824]: E1209 10:06:28.534497 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-webhook-certs podName:e7e496dc-23c4-47cd-9c5a-aa4430130849 nodeName:}" failed. No retries permitted until 2025-12-09 10:06:44.534470584 +0000 UTC m=+1160.868975251 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-webhook-certs") pod "openstack-operator-controller-manager-598df84fd8-l2bt7" (UID: "e7e496dc-23c4-47cd-9c5a-aa4430130849") : secret "webhook-server-cert" not found Dec 09 10:06:29 crc kubenswrapper[4824]: I1209 10:06:29.632854 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-metrics-certs\") pod \"openstack-operator-controller-manager-598df84fd8-l2bt7\" (UID: \"e7e496dc-23c4-47cd-9c5a-aa4430130849\") " pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 10:06:29 crc kubenswrapper[4824]: I1209 10:06:29.701612 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-operator-b846b8d48-dhtnn" podUID="ebb1c8e1-3960-4c38-969b-c483143c87aa" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.101:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 10:06:29 crc kubenswrapper[4824]: I1209 10:06:29.941364 4824 patch_prober.go:28] interesting pod/route-controller-manager-7f84c5b8f9-rqmr9 container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 10:06:29 crc kubenswrapper[4824]: I1209 10:06:29.941428 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" podUID="afe92d2b-bce3-445e-9c26-9c533353e7a3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 10:06:30 crc kubenswrapper[4824]: E1209 10:06:30.039959 4824 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.517s" Dec 09 10:06:32 crc kubenswrapper[4824]: E1209 10:06:32.160978 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea" Dec 09 10:06:32 crc kubenswrapper[4824]: E1209 10:06:32.161871 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lbzv9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-6mfg8_openstack-operators(5b34da01-28d8-476c-a2a5-a098489d90c1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:06:38 crc kubenswrapper[4824]: I1209 10:06:38.295029 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" podUID="9f87669d-33a9-4269-808c-ca6c718f762c" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.44:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 10:06:39 crc kubenswrapper[4824]: E1209 10:06:39.263494 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a" Dec 09 10:06:39 crc kubenswrapper[4824]: E1209 10:06:39.266450 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rxmpl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5b5fd79c9c-9z2kx_openstack-operators(d1323625-eb0f-4606-bee7-7acbd453d4a5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:06:43 crc kubenswrapper[4824]: E1209 10:06:43.007956 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad" Dec 09 10:06:43 crc kubenswrapper[4824]: E1209 10:06:43.008616 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-klpl6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-79c8c4686c-r9fcl_openstack-operators(b6577050-6a21-43bb-84aa-20aad247aafc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:06:44 crc kubenswrapper[4824]: E1209 10:06:44.016528 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 09 10:06:44 crc kubenswrapper[4824]: E1209 10:06:44.017062 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q5sx5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-k92kx_openstack-operators(30f20f09-f33f-4c7a-a4b4-41b51f73d692): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:06:44 crc kubenswrapper[4824]: I1209 10:06:44.626292 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-webhook-certs\") pod \"openstack-operator-controller-manager-598df84fd8-l2bt7\" (UID: \"e7e496dc-23c4-47cd-9c5a-aa4430130849\") " pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 10:06:44 crc kubenswrapper[4824]: I1209 10:06:44.634504 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e7e496dc-23c4-47cd-9c5a-aa4430130849-webhook-certs\") pod \"openstack-operator-controller-manager-598df84fd8-l2bt7\" (UID: \"e7e496dc-23c4-47cd-9c5a-aa4430130849\") " pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 10:06:44 crc kubenswrapper[4824]: E1209 10:06:44.692731 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557" Dec 09 10:06:44 crc kubenswrapper[4824]: E1209 10:06:44.693053 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d577n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-hlwv8_openstack-operators(23621a98-1d6b-421a-976c-965cecfe6db1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:06:44 crc kubenswrapper[4824]: I1209 10:06:44.857701 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-wbthm" Dec 09 10:06:44 crc kubenswrapper[4824]: I1209 10:06:44.866049 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 10:06:54 crc kubenswrapper[4824]: E1209 10:06:54.192115 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027" Dec 09 10:06:54 crc kubenswrapper[4824]: E1209 10:06:54.193233 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5gmrg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-5697bb5779-vx285_openstack-operators(54693584-08f2-4326-8062-92d68899d7e4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:06:54 crc kubenswrapper[4824]: E1209 10:06:54.705343 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:981b6a8f95934a86c5f10ef6e198b07265aeba7f11cf84b9ccd13dfaf06f3ca3" Dec 09 10:06:54 crc kubenswrapper[4824]: E1209 10:06:54.705917 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:981b6a8f95934a86c5f10ef6e198b07265aeba7f11cf84b9ccd13dfaf06f3ca3,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-c2w6b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-6c677c69b-lpf4s_openstack-operators(270a1cf5-f943-4749-8c9a-1c6adf3c45b4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:06:55 crc kubenswrapper[4824]: E1209 10:06:55.205913 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:900050d3501c0785b227db34b89883efe68247816e5c7427cacb74f8aa10605a" Dec 09 10:06:55 crc kubenswrapper[4824]: E1209 10:06:55.206499 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:900050d3501c0785b227db34b89883efe68247816e5c7427cacb74f8aa10605a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ph5x7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-697fb699cf-n45n4_openstack-operators(143e48cd-c956-448c-8dcb-5858a582e29c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:06:59 crc kubenswrapper[4824]: E1209 10:06:59.778420 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8" Dec 09 10:06:59 crc kubenswrapper[4824]: E1209 10:06:59.780757 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lq56g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-667bd8d554-q7psb_openstack-operators(0af0249c-ed71-4f92-b1a3-cb6f43b92529): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:07:00 crc kubenswrapper[4824]: E1209 10:07:00.376567 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 09 10:07:00 crc kubenswrapper[4824]: E1209 10:07:00.377000 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-x7wtx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-bf7pj_openstack-operators(ccc192e8-2648-4b1b-8420-1cbd1a27a916): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:07:03 crc kubenswrapper[4824]: E1209 10:07:03.033918 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 10:07:03 crc kubenswrapper[4824]: E1209 10:07:03.034590 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lbzv9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-6mfg8_openstack-operators(5b34da01-28d8-476c-a2a5-a098489d90c1): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 09 10:07:03 crc kubenswrapper[4824]: E1209 10:07:03.035838 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6mfg8" podUID="5b34da01-28d8-476c-a2a5-a098489d90c1" Dec 09 10:07:04 crc kubenswrapper[4824]: E1209 10:07:04.893543 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991" Dec 09 10:07:04 crc kubenswrapper[4824]: E1209 10:07:04.893749 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b2k8c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9d58d64bc-rz4sl_openstack-operators(6e91c44f-6221-4dfd-9c18-2bbedff02850): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:07:04 crc kubenswrapper[4824]: E1209 10:07:04.906121 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 10:07:04 crc kubenswrapper[4824]: E1209 10:07:04.906323 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rxmpl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5b5fd79c9c-9z2kx_openstack-operators(d1323625-eb0f-4606-bee7-7acbd453d4a5): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 09 10:07:04 crc kubenswrapper[4824]: E1209 10:07:04.906309 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 10:07:04 crc kubenswrapper[4824]: E1209 10:07:04.906428 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-klpl6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-79c8c4686c-r9fcl_openstack-operators(b6577050-6a21-43bb-84aa-20aad247aafc): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 09 10:07:04 crc kubenswrapper[4824]: E1209 10:07:04.909040 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl" podUID="b6577050-6a21-43bb-84aa-20aad247aafc" Dec 09 10:07:04 crc kubenswrapper[4824]: E1209 10:07:04.909063 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9z2kx" podUID="d1323625-eb0f-4606-bee7-7acbd453d4a5" Dec 09 10:07:05 crc kubenswrapper[4824]: E1209 10:07:05.483521 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94" Dec 09 10:07:05 crc kubenswrapper[4824]: E1209 10:07:05.484124 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cwtqd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-55kwm_openstack-operators(bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:07:06 crc kubenswrapper[4824]: E1209 10:07:06.092385 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 09 10:07:06 crc kubenswrapper[4824]: E1209 10:07:06.092661 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-thjwf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-pqhrq_openstack-operators(383a67da-0f83-41e2-82d6-809d1b6cdbec): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:07:06 crc kubenswrapper[4824]: E1209 10:07:06.515766 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 09 10:07:06 crc kubenswrapper[4824]: E1209 10:07:06.516321 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m7zdc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-z68l9_openstack-operators(31749799-9eb4-403f-a61d-9d50d8bc8367): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:07:11 crc kubenswrapper[4824]: E1209 10:07:11.310066 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 09 10:07:11 crc kubenswrapper[4824]: E1209 10:07:11.311156 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f8ql8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-4q2h7_openstack-operators(7de1c7e2-adcb-4311-926c-ffc42e3e5fde): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:07:12 crc kubenswrapper[4824]: E1209 10:07:12.326052 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 09 10:07:12 crc kubenswrapper[4824]: E1209 10:07:12.326281 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qh8s2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-rbwkh_openstack-operators(d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:07:14 crc kubenswrapper[4824]: E1209 10:07:14.958651 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 09 10:07:14 crc kubenswrapper[4824]: E1209 10:07:14.959973 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lmfjn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-lwtpc_openstack-operators(09d4392b-616f-4531-837e-46f1ef396caf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:07:14 crc kubenswrapper[4824]: E1209 10:07:14.961190 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwtpc" podUID="09d4392b-616f-4531-837e-46f1ef396caf" Dec 09 10:07:15 crc kubenswrapper[4824]: E1209 10:07:15.048163 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.20:5001/openstack-k8s-operators/telemetry-operator:adc44e0841ee1c9c853afa02c8f62dffad25da1d" Dec 09 10:07:15 crc kubenswrapper[4824]: E1209 10:07:15.048241 4824 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.20:5001/openstack-k8s-operators/telemetry-operator:adc44e0841ee1c9c853afa02c8f62dffad25da1d" Dec 09 10:07:15 crc kubenswrapper[4824]: E1209 10:07:15.048812 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.20:5001/openstack-k8s-operators/telemetry-operator:adc44e0841ee1c9c853afa02c8f62dffad25da1d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gl2dz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-577c7bd649-xtnkl_openstack-operators(feab57e6-788b-4f23-9e99-aa248786052d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:07:15 crc kubenswrapper[4824]: I1209 10:07:15.509985 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284"] Dec 09 10:07:15 crc kubenswrapper[4824]: I1209 10:07:15.706183 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7"] Dec 09 10:07:15 crc kubenswrapper[4824]: W1209 10:07:15.783594 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7e496dc_23c4_47cd_9c5a_aa4430130849.slice/crio-ac224e1ab794773cbf5c554826b3e5da586b462c3f674ab969e8eb0de173450b WatchSource:0}: Error finding container ac224e1ab794773cbf5c554826b3e5da586b462c3f674ab969e8eb0de173450b: Status 404 returned error can't find the container with id ac224e1ab794773cbf5c554826b3e5da586b462c3f674ab969e8eb0de173450b Dec 09 10:07:15 crc kubenswrapper[4824]: I1209 10:07:15.793752 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg"] Dec 09 10:07:15 crc kubenswrapper[4824]: W1209 10:07:15.924414 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1041e8ed_5d4a_4d51_82fb_b1d9f02e5e97.slice/crio-31c90912863f86486ccb815f3dcdfcf4af9783cbb9c42fc19b7d418301659cdd WatchSource:0}: Error finding container 31c90912863f86486ccb815f3dcdfcf4af9783cbb9c42fc19b7d418301659cdd: Status 404 returned error can't find the container with id 31c90912863f86486ccb815f3dcdfcf4af9783cbb9c42fc19b7d418301659cdd Dec 09 10:07:16 crc kubenswrapper[4824]: I1209 10:07:16.148680 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" event={"ID":"c2889de3-49b2-4465-8235-37f9e58b42a3","Type":"ContainerStarted","Data":"2b7aad79abfc9afa9026a7a11ef659928e3aa364ac8e77a809b473773f22de05"} Dec 09 10:07:16 crc kubenswrapper[4824]: I1209 10:07:16.244239 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" event={"ID":"e7e496dc-23c4-47cd-9c5a-aa4430130849","Type":"ContainerStarted","Data":"ac224e1ab794773cbf5c554826b3e5da586b462c3f674ab969e8eb0de173450b"} Dec 09 10:07:16 crc kubenswrapper[4824]: I1209 10:07:16.250925 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" event={"ID":"1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97","Type":"ContainerStarted","Data":"31c90912863f86486ccb815f3dcdfcf4af9783cbb9c42fc19b7d418301659cdd"} Dec 09 10:07:16 crc kubenswrapper[4824]: I1209 10:07:16.262017 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh" event={"ID":"91bf6cd9-a4fe-46e8-b1e7-39e57c1398b3","Type":"ContainerStarted","Data":"88f94af10301be21fa06f2882918f32b08699007d692a7eaf3f94bdadb4378fe"} Dec 09 10:07:17 crc kubenswrapper[4824]: I1209 10:07:17.272302 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-lfprd" event={"ID":"b0c20e13-b233-4b21-8f47-facef9603735","Type":"ContainerStarted","Data":"715a775b8095fce3f440a6b8d96da8518c365f38a1bcb1424861cf469df16f8a"} Dec 09 10:07:18 crc kubenswrapper[4824]: I1209 10:07:18.302098 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9z2kx" event={"ID":"d1323625-eb0f-4606-bee7-7acbd453d4a5","Type":"ContainerStarted","Data":"744840f111ef55f13e1ed6f08378140492c8cc75e2a1053fb0cd8a00748b7dbe"} Dec 09 10:07:18 crc kubenswrapper[4824]: I1209 10:07:18.309475 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl" event={"ID":"b6577050-6a21-43bb-84aa-20aad247aafc","Type":"ContainerStarted","Data":"1e3e8cc3634bbda54fc7400d8c9ba593261a4d358dd2d6f34df27727570ed602"} Dec 09 10:07:18 crc kubenswrapper[4824]: I1209 10:07:18.313715 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6mfg8" event={"ID":"5b34da01-28d8-476c-a2a5-a098489d90c1","Type":"ContainerStarted","Data":"5fad51c856bf482fe312f0b43cf6ae2559ef60c940c650c18986e3eeabf662dc"} Dec 09 10:07:19 crc kubenswrapper[4824]: I1209 10:07:19.323860 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" event={"ID":"e7e496dc-23c4-47cd-9c5a-aa4430130849","Type":"ContainerStarted","Data":"77ee6a4f989b6c4685d5daaf4f923c657aa4f8120f4031464ece9cf6a217d1b7"} Dec 09 10:07:19 crc kubenswrapper[4824]: I1209 10:07:19.324324 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 10:07:19 crc kubenswrapper[4824]: I1209 10:07:19.366512 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" podStartSLOduration=68.366468388 podStartE2EDuration="1m8.366468388s" podCreationTimestamp="2025-12-09 10:06:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:07:19.357630179 +0000 UTC m=+1195.692134846" watchObservedRunningTime="2025-12-09 10:07:19.366468388 +0000 UTC m=+1195.700973055" Dec 09 10:07:19 crc kubenswrapper[4824]: E1209 10:07:19.579336 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-bf7pj" podUID="ccc192e8-2648-4b1b-8420-1cbd1a27a916" Dec 09 10:07:19 crc kubenswrapper[4824]: E1209 10:07:19.607914 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-k92kx" podUID="30f20f09-f33f-4c7a-a4b4-41b51f73d692" Dec 09 10:07:19 crc kubenswrapper[4824]: E1209 10:07:19.722892 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-q7psb" podUID="0af0249c-ed71-4f92-b1a3-cb6f43b92529" Dec 09 10:07:19 crc kubenswrapper[4824]: E1209 10:07:19.723414 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4" podUID="143e48cd-c956-448c-8dcb-5858a582e29c" Dec 09 10:07:20 crc kubenswrapper[4824]: E1209 10:07:20.201072 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" podUID="bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8" Dec 09 10:07:20 crc kubenswrapper[4824]: E1209 10:07:20.227226 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vx285" podUID="54693584-08f2-4326-8062-92d68899d7e4" Dec 09 10:07:20 crc kubenswrapper[4824]: I1209 10:07:20.341019 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" event={"ID":"bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8","Type":"ContainerStarted","Data":"2645752d9c54c5ec15fd61fa884d848e74a2a7ddfc3e01f0b95940170ce17c85"} Dec 09 10:07:20 crc kubenswrapper[4824]: I1209 10:07:20.343939 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vx285" event={"ID":"54693584-08f2-4326-8062-92d68899d7e4","Type":"ContainerStarted","Data":"67d514b54203a235d2124a54c61b5dff42460418173d99d3a0b179d0ef4cb2f2"} Dec 09 10:07:20 crc kubenswrapper[4824]: E1209 10:07:20.355359 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" podUID="bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8" Dec 09 10:07:20 crc kubenswrapper[4824]: I1209 10:07:20.360489 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-k92kx" event={"ID":"30f20f09-f33f-4c7a-a4b4-41b51f73d692","Type":"ContainerStarted","Data":"07ccf78feacc3cfbfd522268e2f0bfe766be7fb284fdee58af0ed8b551bce936"} Dec 09 10:07:20 crc kubenswrapper[4824]: I1209 10:07:20.368337 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6mfg8" event={"ID":"5b34da01-28d8-476c-a2a5-a098489d90c1","Type":"ContainerStarted","Data":"0183380cf8625008680f4a286f4dc997fa4862001e469083992b1b4837b835cb"} Dec 09 10:07:20 crc kubenswrapper[4824]: I1209 10:07:20.368413 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6mfg8" Dec 09 10:07:20 crc kubenswrapper[4824]: I1209 10:07:20.378543 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-bf7pj" event={"ID":"ccc192e8-2648-4b1b-8420-1cbd1a27a916","Type":"ContainerStarted","Data":"888f9f5fd08dca8b2269bedd84f93ef2e55ce1bd79ee223a959c4cdc87b8f3e6"} Dec 09 10:07:20 crc kubenswrapper[4824]: I1209 10:07:20.384863 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl" event={"ID":"b6577050-6a21-43bb-84aa-20aad247aafc","Type":"ContainerStarted","Data":"2fda835a7fa01304115eee5aee4fe7b0bb4fc2063a45004a8307e1bca0f8c19e"} Dec 09 10:07:20 crc kubenswrapper[4824]: I1209 10:07:20.385930 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl" Dec 09 10:07:20 crc kubenswrapper[4824]: I1209 10:07:20.398871 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-q7psb" event={"ID":"0af0249c-ed71-4f92-b1a3-cb6f43b92529","Type":"ContainerStarted","Data":"4d6a56917a845d14d4bb23914044038a3acf58afac371219e6cd9da841666522"} Dec 09 10:07:20 crc kubenswrapper[4824]: I1209 10:07:20.412684 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4" event={"ID":"143e48cd-c956-448c-8dcb-5858a582e29c","Type":"ContainerStarted","Data":"ce1aa087ffb8ca6e869a8e6c11a38dc530e7a1f9653c4317b2729794edfb33de"} Dec 09 10:07:20 crc kubenswrapper[4824]: I1209 10:07:20.422370 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-lfprd" event={"ID":"b0c20e13-b233-4b21-8f47-facef9603735","Type":"ContainerStarted","Data":"9f75cf64249ebd1edc041db0b6eb2a786667b1cd190ddd1cf881cf26d91a69ce"} Dec 09 10:07:20 crc kubenswrapper[4824]: I1209 10:07:20.422421 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-967d97867-lfprd" Dec 09 10:07:20 crc kubenswrapper[4824]: I1209 10:07:20.448370 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6mfg8" podStartSLOduration=3.627534155 podStartE2EDuration="1m10.448347394s" podCreationTimestamp="2025-12-09 10:06:10 +0000 UTC" firstStartedPulling="2025-12-09 10:06:12.400902401 +0000 UTC m=+1128.735407068" lastFinishedPulling="2025-12-09 10:07:19.22171563 +0000 UTC m=+1195.556220307" observedRunningTime="2025-12-09 10:07:20.441190309 +0000 UTC m=+1196.775694976" watchObservedRunningTime="2025-12-09 10:07:20.448347394 +0000 UTC m=+1196.782852061" Dec 09 10:07:20 crc kubenswrapper[4824]: I1209 10:07:20.474931 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl" podStartSLOduration=6.391029728 podStartE2EDuration="1m10.474911103s" podCreationTimestamp="2025-12-09 10:06:10 +0000 UTC" firstStartedPulling="2025-12-09 10:06:15.342208625 +0000 UTC m=+1131.676713292" lastFinishedPulling="2025-12-09 10:07:19.42609 +0000 UTC m=+1195.760594667" observedRunningTime="2025-12-09 10:07:20.47481222 +0000 UTC m=+1196.809316897" watchObservedRunningTime="2025-12-09 10:07:20.474911103 +0000 UTC m=+1196.809415780" Dec 09 10:07:20 crc kubenswrapper[4824]: I1209 10:07:20.585410 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-967d97867-lfprd" podStartSLOduration=6.759984145 podStartE2EDuration="1m10.585383331s" podCreationTimestamp="2025-12-09 10:06:10 +0000 UTC" firstStartedPulling="2025-12-09 10:06:15.484725974 +0000 UTC m=+1131.819230641" lastFinishedPulling="2025-12-09 10:07:19.31012516 +0000 UTC m=+1195.644629827" observedRunningTime="2025-12-09 10:07:20.579116813 +0000 UTC m=+1196.913621490" watchObservedRunningTime="2025-12-09 10:07:20.585383331 +0000 UTC m=+1196.919888008" Dec 09 10:07:20 crc kubenswrapper[4824]: E1209 10:07:20.672628 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9" podUID="31749799-9eb4-403f-a61d-9d50d8bc8367" Dec 09 10:07:20 crc kubenswrapper[4824]: E1209 10:07:20.811658 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8" podUID="23621a98-1d6b-421a-976c-965cecfe6db1" Dec 09 10:07:20 crc kubenswrapper[4824]: E1209 10:07:20.816379 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" podUID="feab57e6-788b-4f23-9e99-aa248786052d" Dec 09 10:07:20 crc kubenswrapper[4824]: E1209 10:07:20.819466 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" podUID="383a67da-0f83-41e2-82d6-809d1b6cdbec" Dec 09 10:07:21 crc kubenswrapper[4824]: E1209 10:07:21.083400 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" podUID="d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e" Dec 09 10:07:21 crc kubenswrapper[4824]: E1209 10:07:21.084080 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4q2h7" podUID="7de1c7e2-adcb-4311-926c-ffc42e3e5fde" Dec 09 10:07:21 crc kubenswrapper[4824]: E1209 10:07:21.100143 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-lpf4s" podUID="270a1cf5-f943-4749-8c9a-1c6adf3c45b4" Dec 09 10:07:21 crc kubenswrapper[4824]: E1209 10:07:21.102101 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl" podUID="6e91c44f-6221-4dfd-9c18-2bbedff02850" Dec 09 10:07:21 crc kubenswrapper[4824]: I1209 10:07:21.210920 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-967d97867-lfprd" Dec 09 10:07:21 crc kubenswrapper[4824]: I1209 10:07:21.434039 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl" event={"ID":"6e91c44f-6221-4dfd-9c18-2bbedff02850","Type":"ContainerStarted","Data":"60e5a4e8da3f06ef071558053758a4a96685ca9749f83561edf24cb1a8a36c5b"} Dec 09 10:07:21 crc kubenswrapper[4824]: I1209 10:07:21.438746 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9" event={"ID":"31749799-9eb4-403f-a61d-9d50d8bc8367","Type":"ContainerStarted","Data":"5da72be3be8ae0ee5f12fd77e33769d46ba05692a40d85f8f439089df276d4e4"} Dec 09 10:07:21 crc kubenswrapper[4824]: E1209 10:07:21.440591 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9" podUID="31749799-9eb4-403f-a61d-9d50d8bc8367" Dec 09 10:07:21 crc kubenswrapper[4824]: I1209 10:07:21.457228 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8" event={"ID":"23621a98-1d6b-421a-976c-965cecfe6db1","Type":"ContainerStarted","Data":"4a5518e1bc64278ae4e077d8d5691c97183b187de767e047a0da070c5c07b2ec"} Dec 09 10:07:21 crc kubenswrapper[4824]: I1209 10:07:21.463729 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-lpf4s" event={"ID":"270a1cf5-f943-4749-8c9a-1c6adf3c45b4","Type":"ContainerStarted","Data":"98bbdbd34b0ca7495fb5665a07515aeaae0a24fa0eb2a47324d0602031dff44f"} Dec 09 10:07:21 crc kubenswrapper[4824]: I1209 10:07:21.466023 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" event={"ID":"d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e","Type":"ContainerStarted","Data":"44901f8dc25249c6cea7de4138eb988d054c36455c3f2f333397809ee9071f34"} Dec 09 10:07:21 crc kubenswrapper[4824]: E1209 10:07:21.468762 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" podUID="d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e" Dec 09 10:07:21 crc kubenswrapper[4824]: I1209 10:07:21.488119 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4q2h7" event={"ID":"7de1c7e2-adcb-4311-926c-ffc42e3e5fde","Type":"ContainerStarted","Data":"740a277d03c341ff66002ce036409730319eef7e9a1d17ffbf05bd95dea54c3e"} Dec 09 10:07:21 crc kubenswrapper[4824]: I1209 10:07:21.516653 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" event={"ID":"feab57e6-788b-4f23-9e99-aa248786052d","Type":"ContainerStarted","Data":"8b82903c2c25b2ead75d43bbac88a413e5b364def686705ae3a11b2b47593b7c"} Dec 09 10:07:21 crc kubenswrapper[4824]: E1209 10:07:21.518270 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.20:5001/openstack-k8s-operators/telemetry-operator:adc44e0841ee1c9c853afa02c8f62dffad25da1d\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" podUID="feab57e6-788b-4f23-9e99-aa248786052d" Dec 09 10:07:21 crc kubenswrapper[4824]: I1209 10:07:21.519392 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9z2kx" event={"ID":"d1323625-eb0f-4606-bee7-7acbd453d4a5","Type":"ContainerStarted","Data":"f5904033917cd44585f5b0849c2ba715133c6258e981f54b75b461fc65201b4a"} Dec 09 10:07:21 crc kubenswrapper[4824]: I1209 10:07:21.520449 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9z2kx" Dec 09 10:07:21 crc kubenswrapper[4824]: I1209 10:07:21.539318 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh" event={"ID":"91bf6cd9-a4fe-46e8-b1e7-39e57c1398b3","Type":"ContainerStarted","Data":"771827e0d6a0fa1fff860db9aad6b7fcba7805de79707db01fd681d760728fd9"} Dec 09 10:07:21 crc kubenswrapper[4824]: I1209 10:07:21.539926 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh" Dec 09 10:07:21 crc kubenswrapper[4824]: I1209 10:07:21.556811 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh" Dec 09 10:07:21 crc kubenswrapper[4824]: I1209 10:07:21.565313 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" event={"ID":"383a67da-0f83-41e2-82d6-809d1b6cdbec","Type":"ContainerStarted","Data":"310a38f3cad63f21325d128c1c18d4eb1783d2320e2d294258e9265e9b037e4d"} Dec 09 10:07:21 crc kubenswrapper[4824]: E1209 10:07:21.567259 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" podUID="383a67da-0f83-41e2-82d6-809d1b6cdbec" Dec 09 10:07:21 crc kubenswrapper[4824]: I1209 10:07:21.631921 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh" podStartSLOduration=8.032206499 podStartE2EDuration="1m11.631899421s" podCreationTimestamp="2025-12-09 10:06:10 +0000 UTC" firstStartedPulling="2025-12-09 10:06:16.096353878 +0000 UTC m=+1132.430858545" lastFinishedPulling="2025-12-09 10:07:19.6960468 +0000 UTC m=+1196.030551467" observedRunningTime="2025-12-09 10:07:21.606469998 +0000 UTC m=+1197.940974675" watchObservedRunningTime="2025-12-09 10:07:21.631899421 +0000 UTC m=+1197.966404088" Dec 09 10:07:21 crc kubenswrapper[4824]: I1209 10:07:21.657506 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9z2kx" podStartSLOduration=7.017233753 podStartE2EDuration="1m11.657475548s" podCreationTimestamp="2025-12-09 10:06:10 +0000 UTC" firstStartedPulling="2025-12-09 10:06:15.345552671 +0000 UTC m=+1131.680057338" lastFinishedPulling="2025-12-09 10:07:19.985794466 +0000 UTC m=+1196.320299133" observedRunningTime="2025-12-09 10:07:21.633318655 +0000 UTC m=+1197.967823332" watchObservedRunningTime="2025-12-09 10:07:21.657475548 +0000 UTC m=+1197.991980215" Dec 09 10:07:22 crc kubenswrapper[4824]: I1209 10:07:22.586936 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-k92kx" event={"ID":"30f20f09-f33f-4c7a-a4b4-41b51f73d692","Type":"ContainerStarted","Data":"18eef9d35f7840b22630760c0ad536ed830a0535a6304c1f88fc9e358930411c"} Dec 09 10:07:22 crc kubenswrapper[4824]: I1209 10:07:22.587115 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-k92kx" Dec 09 10:07:22 crc kubenswrapper[4824]: I1209 10:07:22.596081 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-q7psb" event={"ID":"0af0249c-ed71-4f92-b1a3-cb6f43b92529","Type":"ContainerStarted","Data":"582beb4c6ba000dff1857ea81a561fc444ed9f1574280a492dd1fe4a8ece0cdb"} Dec 09 10:07:22 crc kubenswrapper[4824]: I1209 10:07:22.597162 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-q7psb" Dec 09 10:07:22 crc kubenswrapper[4824]: I1209 10:07:22.604618 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-bf7pj" event={"ID":"ccc192e8-2648-4b1b-8420-1cbd1a27a916","Type":"ContainerStarted","Data":"96c5ae3a389a995de757a37c7f98552731203e6c03ea500c2fb6b1f5eaedb77a"} Dec 09 10:07:22 crc kubenswrapper[4824]: I1209 10:07:22.605353 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-bf7pj" Dec 09 10:07:22 crc kubenswrapper[4824]: I1209 10:07:22.621410 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4" event={"ID":"143e48cd-c956-448c-8dcb-5858a582e29c","Type":"ContainerStarted","Data":"7681e32f76e067007f5b25ca0d74b0e725c9aaaee11a16e7e0b9b846235bb2a5"} Dec 09 10:07:22 crc kubenswrapper[4824]: I1209 10:07:22.622005 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4" Dec 09 10:07:22 crc kubenswrapper[4824]: I1209 10:07:22.636164 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-k92kx" podStartSLOduration=6.854518038 podStartE2EDuration="1m12.636136717s" podCreationTimestamp="2025-12-09 10:06:10 +0000 UTC" firstStartedPulling="2025-12-09 10:06:15.399410491 +0000 UTC m=+1131.733915158" lastFinishedPulling="2025-12-09 10:07:21.18102917 +0000 UTC m=+1197.515533837" observedRunningTime="2025-12-09 10:07:22.6267587 +0000 UTC m=+1198.961263387" watchObservedRunningTime="2025-12-09 10:07:22.636136717 +0000 UTC m=+1198.970641374" Dec 09 10:07:22 crc kubenswrapper[4824]: I1209 10:07:22.637289 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vx285" event={"ID":"54693584-08f2-4326-8062-92d68899d7e4","Type":"ContainerStarted","Data":"e87ac27490df1b051f81a8b01a3853352f2e15d2a85307f560d92a1fa022dea9"} Dec 09 10:07:22 crc kubenswrapper[4824]: I1209 10:07:22.637348 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vx285" Dec 09 10:07:22 crc kubenswrapper[4824]: I1209 10:07:22.664655 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 10:07:22 crc kubenswrapper[4824]: I1209 10:07:22.674128 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-q7psb" podStartSLOduration=6.490434138 podStartE2EDuration="1m11.674102484s" podCreationTimestamp="2025-12-09 10:06:11 +0000 UTC" firstStartedPulling="2025-12-09 10:06:16.038593645 +0000 UTC m=+1132.373098322" lastFinishedPulling="2025-12-09 10:07:21.222262001 +0000 UTC m=+1197.556766668" observedRunningTime="2025-12-09 10:07:22.664004056 +0000 UTC m=+1198.998508723" watchObservedRunningTime="2025-12-09 10:07:22.674102484 +0000 UTC m=+1199.008607151" Dec 09 10:07:22 crc kubenswrapper[4824]: I1209 10:07:22.699848 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4" podStartSLOduration=6.634318357 podStartE2EDuration="1m12.699824926s" podCreationTimestamp="2025-12-09 10:06:10 +0000 UTC" firstStartedPulling="2025-12-09 10:06:15.319918842 +0000 UTC m=+1131.654423509" lastFinishedPulling="2025-12-09 10:07:21.385425411 +0000 UTC m=+1197.719930078" observedRunningTime="2025-12-09 10:07:22.688272402 +0000 UTC m=+1199.022777069" watchObservedRunningTime="2025-12-09 10:07:22.699824926 +0000 UTC m=+1199.034329603" Dec 09 10:07:22 crc kubenswrapper[4824]: I1209 10:07:22.715397 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-bf7pj" podStartSLOduration=6.824741558 podStartE2EDuration="1m12.715376798s" podCreationTimestamp="2025-12-09 10:06:10 +0000 UTC" firstStartedPulling="2025-12-09 10:06:15.334614656 +0000 UTC m=+1131.669119323" lastFinishedPulling="2025-12-09 10:07:21.225249896 +0000 UTC m=+1197.559754563" observedRunningTime="2025-12-09 10:07:22.706502858 +0000 UTC m=+1199.041007545" watchObservedRunningTime="2025-12-09 10:07:22.715376798 +0000 UTC m=+1199.049881465" Dec 09 10:07:22 crc kubenswrapper[4824]: I1209 10:07:22.759682 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vx285" podStartSLOduration=6.913047595 podStartE2EDuration="1m12.759659135s" podCreationTimestamp="2025-12-09 10:06:10 +0000 UTC" firstStartedPulling="2025-12-09 10:06:15.334383989 +0000 UTC m=+1131.668888646" lastFinishedPulling="2025-12-09 10:07:21.180995519 +0000 UTC m=+1197.515500186" observedRunningTime="2025-12-09 10:07:22.754638447 +0000 UTC m=+1199.089143124" watchObservedRunningTime="2025-12-09 10:07:22.759659135 +0000 UTC m=+1199.094163802" Dec 09 10:07:23 crc kubenswrapper[4824]: I1209 10:07:23.667409 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-lpf4s" event={"ID":"270a1cf5-f943-4749-8c9a-1c6adf3c45b4","Type":"ContainerStarted","Data":"88d68d15e499a4720522379e3ad95ce00b87f8a0b3d93bafbd60b49a42d9b7a6"} Dec 09 10:07:23 crc kubenswrapper[4824]: I1209 10:07:23.667605 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-lpf4s" Dec 09 10:07:23 crc kubenswrapper[4824]: I1209 10:07:23.671283 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4q2h7" event={"ID":"7de1c7e2-adcb-4311-926c-ffc42e3e5fde","Type":"ContainerStarted","Data":"bf6eac2af9b34ed5e8143ec9a4b5e1f1d2a1cf29d103f12c1fa230c24b2a9b8a"} Dec 09 10:07:23 crc kubenswrapper[4824]: I1209 10:07:23.671409 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4q2h7" Dec 09 10:07:23 crc kubenswrapper[4824]: I1209 10:07:23.677251 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl" event={"ID":"6e91c44f-6221-4dfd-9c18-2bbedff02850","Type":"ContainerStarted","Data":"f0b3f37bdad3f939e038e3d3ace05555dbd1fba79e301122328cdedead3df512"} Dec 09 10:07:23 crc kubenswrapper[4824]: I1209 10:07:23.677446 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl" Dec 09 10:07:23 crc kubenswrapper[4824]: I1209 10:07:23.679599 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8" event={"ID":"23621a98-1d6b-421a-976c-965cecfe6db1","Type":"ContainerStarted","Data":"19c2f889a724f4dacee8917c6176ccbbc55dbfa1bb4f60e0ce6840d832a83568"} Dec 09 10:07:23 crc kubenswrapper[4824]: I1209 10:07:23.683662 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9z2kx" Dec 09 10:07:23 crc kubenswrapper[4824]: I1209 10:07:23.707586 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-lpf4s" podStartSLOduration=6.786275443 podStartE2EDuration="1m13.707562733s" podCreationTimestamp="2025-12-09 10:06:10 +0000 UTC" firstStartedPulling="2025-12-09 10:06:15.283371528 +0000 UTC m=+1131.617876195" lastFinishedPulling="2025-12-09 10:07:22.204658818 +0000 UTC m=+1198.539163485" observedRunningTime="2025-12-09 10:07:23.687051446 +0000 UTC m=+1200.021556123" watchObservedRunningTime="2025-12-09 10:07:23.707562733 +0000 UTC m=+1200.042067400" Dec 09 10:07:23 crc kubenswrapper[4824]: I1209 10:07:23.716697 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4q2h7" podStartSLOduration=6.727741688 podStartE2EDuration="1m13.716672531s" podCreationTimestamp="2025-12-09 10:06:10 +0000 UTC" firstStartedPulling="2025-12-09 10:06:15.335106372 +0000 UTC m=+1131.669611039" lastFinishedPulling="2025-12-09 10:07:22.324037215 +0000 UTC m=+1198.658541882" observedRunningTime="2025-12-09 10:07:23.704463465 +0000 UTC m=+1200.038968142" watchObservedRunningTime="2025-12-09 10:07:23.716672531 +0000 UTC m=+1200.051177188" Dec 09 10:07:23 crc kubenswrapper[4824]: I1209 10:07:23.726454 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl" podStartSLOduration=7.475542528 podStartE2EDuration="1m13.726434159s" podCreationTimestamp="2025-12-09 10:06:10 +0000 UTC" firstStartedPulling="2025-12-09 10:06:16.058945667 +0000 UTC m=+1132.393450334" lastFinishedPulling="2025-12-09 10:07:22.309837308 +0000 UTC m=+1198.644341965" observedRunningTime="2025-12-09 10:07:23.723228698 +0000 UTC m=+1200.057733365" watchObservedRunningTime="2025-12-09 10:07:23.726434159 +0000 UTC m=+1200.060938826" Dec 09 10:07:23 crc kubenswrapper[4824]: I1209 10:07:23.788394 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8" podStartSLOduration=7.501128716 podStartE2EDuration="1m13.788375123s" podCreationTimestamp="2025-12-09 10:06:10 +0000 UTC" firstStartedPulling="2025-12-09 10:06:16.036782488 +0000 UTC m=+1132.371287155" lastFinishedPulling="2025-12-09 10:07:22.324028895 +0000 UTC m=+1198.658533562" observedRunningTime="2025-12-09 10:07:23.786924228 +0000 UTC m=+1200.121428905" watchObservedRunningTime="2025-12-09 10:07:23.788375123 +0000 UTC m=+1200.122879791" Dec 09 10:07:24 crc kubenswrapper[4824]: I1209 10:07:24.689907 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8" Dec 09 10:07:24 crc kubenswrapper[4824]: I1209 10:07:24.875740 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 10:07:25 crc kubenswrapper[4824]: I1209 10:07:25.700375 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" event={"ID":"d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e","Type":"ContainerStarted","Data":"2c66a87b3d637d530e0d9f96512da5a19cae973cc6c8de51cc3f2c64838db824"} Dec 09 10:07:25 crc kubenswrapper[4824]: I1209 10:07:25.700969 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" Dec 09 10:07:25 crc kubenswrapper[4824]: I1209 10:07:25.702040 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" event={"ID":"c2889de3-49b2-4465-8235-37f9e58b42a3","Type":"ContainerStarted","Data":"c4a6bd38a195f4eb3c912c97bc78383959a4e5d35e8611606e390dd61f32d99e"} Dec 09 10:07:25 crc kubenswrapper[4824]: I1209 10:07:25.704247 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" event={"ID":"1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97","Type":"ContainerStarted","Data":"3d115f53f24fc30a6f79e8e6af49df8f72cdc2914c63e1abf2df76f7c30a5b07"} Dec 09 10:07:25 crc kubenswrapper[4824]: I1209 10:07:25.704300 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" event={"ID":"1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97","Type":"ContainerStarted","Data":"d0f63e8e47831b2e597ad2f23667a503826a77f9276c5883b6c04777ccf6d075"} Dec 09 10:07:25 crc kubenswrapper[4824]: I1209 10:07:25.704492 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" Dec 09 10:07:25 crc kubenswrapper[4824]: I1209 10:07:25.726012 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" podStartSLOduration=6.629900877 podStartE2EDuration="1m15.725988449s" podCreationTimestamp="2025-12-09 10:06:10 +0000 UTC" firstStartedPulling="2025-12-09 10:06:16.038580024 +0000 UTC m=+1132.373084701" lastFinishedPulling="2025-12-09 10:07:25.134667606 +0000 UTC m=+1201.469172273" observedRunningTime="2025-12-09 10:07:25.723887093 +0000 UTC m=+1202.058391770" watchObservedRunningTime="2025-12-09 10:07:25.725988449 +0000 UTC m=+1202.060493116" Dec 09 10:07:25 crc kubenswrapper[4824]: I1209 10:07:25.757503 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" podStartSLOduration=66.54294158 podStartE2EDuration="1m15.757477923s" podCreationTimestamp="2025-12-09 10:06:10 +0000 UTC" firstStartedPulling="2025-12-09 10:07:15.927405853 +0000 UTC m=+1192.261910520" lastFinishedPulling="2025-12-09 10:07:25.141942186 +0000 UTC m=+1201.476446863" observedRunningTime="2025-12-09 10:07:25.754123907 +0000 UTC m=+1202.088628584" watchObservedRunningTime="2025-12-09 10:07:25.757477923 +0000 UTC m=+1202.091982590" Dec 09 10:07:25 crc kubenswrapper[4824]: E1209 10:07:25.917090 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwtpc" podUID="09d4392b-616f-4531-837e-46f1ef396caf" Dec 09 10:07:26 crc kubenswrapper[4824]: I1209 10:07:26.715264 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" event={"ID":"c2889de3-49b2-4465-8235-37f9e58b42a3","Type":"ContainerStarted","Data":"2302c10a85804e8e05326004fcf0fc49a95ff3ac6b453cd3a88971164934f5c7"} Dec 09 10:07:26 crc kubenswrapper[4824]: I1209 10:07:26.716888 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" Dec 09 10:07:26 crc kubenswrapper[4824]: I1209 10:07:26.750887 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" podStartSLOduration=67.40355765300001 podStartE2EDuration="1m16.750589498s" podCreationTimestamp="2025-12-09 10:06:10 +0000 UTC" firstStartedPulling="2025-12-09 10:07:15.787381573 +0000 UTC m=+1192.121886240" lastFinishedPulling="2025-12-09 10:07:25.134413418 +0000 UTC m=+1201.468918085" observedRunningTime="2025-12-09 10:07:26.742254835 +0000 UTC m=+1203.076759522" watchObservedRunningTime="2025-12-09 10:07:26.750589498 +0000 UTC m=+1203.085094175" Dec 09 10:07:30 crc kubenswrapper[4824]: I1209 10:07:30.797455 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6mfg8" Dec 09 10:07:30 crc kubenswrapper[4824]: I1209 10:07:30.816678 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-lpf4s" Dec 09 10:07:30 crc kubenswrapper[4824]: I1209 10:07:30.882759 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4" Dec 09 10:07:30 crc kubenswrapper[4824]: I1209 10:07:30.901423 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vx285" Dec 09 10:07:30 crc kubenswrapper[4824]: I1209 10:07:30.996313 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-bf7pj" Dec 09 10:07:30 crc kubenswrapper[4824]: I1209 10:07:30.996390 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-k92kx" Dec 09 10:07:31 crc kubenswrapper[4824]: I1209 10:07:31.255340 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4q2h7" Dec 09 10:07:31 crc kubenswrapper[4824]: I1209 10:07:31.489756 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl" Dec 09 10:07:31 crc kubenswrapper[4824]: I1209 10:07:31.836037 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8" Dec 09 10:07:31 crc kubenswrapper[4824]: I1209 10:07:31.865958 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" Dec 09 10:07:31 crc kubenswrapper[4824]: I1209 10:07:31.973223 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl" Dec 09 10:07:32 crc kubenswrapper[4824]: I1209 10:07:32.111659 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-q7psb" Dec 09 10:07:32 crc kubenswrapper[4824]: I1209 10:07:32.861640 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:07:32 crc kubenswrapper[4824]: I1209 10:07:32.861717 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:07:34 crc kubenswrapper[4824]: I1209 10:07:34.786419 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9" event={"ID":"31749799-9eb4-403f-a61d-9d50d8bc8367","Type":"ContainerStarted","Data":"e93d0d1a455f3e00e021ca3fe992e4a17bb89a9c336c679aba1108dac4229fd3"} Dec 09 10:07:34 crc kubenswrapper[4824]: I1209 10:07:34.787343 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9" Dec 09 10:07:34 crc kubenswrapper[4824]: I1209 10:07:34.810477 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9" podStartSLOduration=6.630776336 podStartE2EDuration="1m24.810451576s" podCreationTimestamp="2025-12-09 10:06:10 +0000 UTC" firstStartedPulling="2025-12-09 10:06:16.217140721 +0000 UTC m=+1132.551645398" lastFinishedPulling="2025-12-09 10:07:34.396815971 +0000 UTC m=+1210.731320638" observedRunningTime="2025-12-09 10:07:34.805379896 +0000 UTC m=+1211.139884583" watchObservedRunningTime="2025-12-09 10:07:34.810451576 +0000 UTC m=+1211.144956243" Dec 09 10:07:34 crc kubenswrapper[4824]: E1209 10:07:34.912478 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.20:5001/openstack-k8s-operators/telemetry-operator:adc44e0841ee1c9c853afa02c8f62dffad25da1d\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" podUID="feab57e6-788b-4f23-9e99-aa248786052d" Dec 09 10:07:35 crc kubenswrapper[4824]: I1209 10:07:35.796231 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" event={"ID":"bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8","Type":"ContainerStarted","Data":"4aecf4910f255e94b17caaa7fae8c9aaa0905e7e1ce503d990addff287a7a433"} Dec 09 10:07:35 crc kubenswrapper[4824]: I1209 10:07:35.796823 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" Dec 09 10:07:35 crc kubenswrapper[4824]: I1209 10:07:35.813169 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" podStartSLOduration=5.608967176 podStartE2EDuration="1m24.813149243s" podCreationTimestamp="2025-12-09 10:06:11 +0000 UTC" firstStartedPulling="2025-12-09 10:06:16.215298352 +0000 UTC m=+1132.549803019" lastFinishedPulling="2025-12-09 10:07:35.419480419 +0000 UTC m=+1211.753985086" observedRunningTime="2025-12-09 10:07:35.811616346 +0000 UTC m=+1212.146121013" watchObservedRunningTime="2025-12-09 10:07:35.813149243 +0000 UTC m=+1212.147653910" Dec 09 10:07:36 crc kubenswrapper[4824]: I1209 10:07:36.986394 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" Dec 09 10:07:37 crc kubenswrapper[4824]: I1209 10:07:37.555162 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" Dec 09 10:07:37 crc kubenswrapper[4824]: I1209 10:07:37.816387 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" event={"ID":"383a67da-0f83-41e2-82d6-809d1b6cdbec","Type":"ContainerStarted","Data":"9ea2c6d7fa34ab9c9c15a6234179926d1216a3276cfdc22d1a4626a6ec0c4f5d"} Dec 09 10:07:37 crc kubenswrapper[4824]: I1209 10:07:37.816664 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" Dec 09 10:07:37 crc kubenswrapper[4824]: I1209 10:07:37.837378 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" podStartSLOduration=6.612710646 podStartE2EDuration="1m27.837357402s" podCreationTimestamp="2025-12-09 10:06:10 +0000 UTC" firstStartedPulling="2025-12-09 10:06:16.147284996 +0000 UTC m=+1132.481789663" lastFinishedPulling="2025-12-09 10:07:37.371931752 +0000 UTC m=+1213.706436419" observedRunningTime="2025-12-09 10:07:37.833205681 +0000 UTC m=+1214.167710368" watchObservedRunningTime="2025-12-09 10:07:37.837357402 +0000 UTC m=+1214.171862069" Dec 09 10:07:38 crc kubenswrapper[4824]: I1209 10:07:38.826532 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwtpc" event={"ID":"09d4392b-616f-4531-837e-46f1ef396caf","Type":"ContainerStarted","Data":"fabb8c6131319c5ac951723dc37e441e38f465f708e361fac07da55ae9bf17c4"} Dec 09 10:07:38 crc kubenswrapper[4824]: I1209 10:07:38.844281 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lwtpc" podStartSLOduration=6.470156889 podStartE2EDuration="1m27.844255422s" podCreationTimestamp="2025-12-09 10:06:11 +0000 UTC" firstStartedPulling="2025-12-09 10:06:16.236914025 +0000 UTC m=+1132.571418692" lastFinishedPulling="2025-12-09 10:07:37.611012558 +0000 UTC m=+1213.945517225" observedRunningTime="2025-12-09 10:07:38.84007258 +0000 UTC m=+1215.174577247" watchObservedRunningTime="2025-12-09 10:07:38.844255422 +0000 UTC m=+1215.178760089" Dec 09 10:07:41 crc kubenswrapper[4824]: I1209 10:07:41.971602 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9" Dec 09 10:07:42 crc kubenswrapper[4824]: I1209 10:07:42.095076 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" Dec 09 10:07:47 crc kubenswrapper[4824]: I1209 10:07:47.899552 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" event={"ID":"feab57e6-788b-4f23-9e99-aa248786052d","Type":"ContainerStarted","Data":"fe509de64aa6046a4afdd744d7b9a144392cce3237541624889199bb17b79b7b"} Dec 09 10:07:47 crc kubenswrapper[4824]: I1209 10:07:47.900387 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" Dec 09 10:07:47 crc kubenswrapper[4824]: I1209 10:07:47.923035 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" podStartSLOduration=6.967294138 podStartE2EDuration="1m37.923012919s" podCreationTimestamp="2025-12-09 10:06:10 +0000 UTC" firstStartedPulling="2025-12-09 10:06:16.199618208 +0000 UTC m=+1132.534122875" lastFinishedPulling="2025-12-09 10:07:47.155336989 +0000 UTC m=+1223.489841656" observedRunningTime="2025-12-09 10:07:47.917935258 +0000 UTC m=+1224.252439925" watchObservedRunningTime="2025-12-09 10:07:47.923012919 +0000 UTC m=+1224.257517586" Dec 09 10:07:51 crc kubenswrapper[4824]: I1209 10:07:51.902003 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" Dec 09 10:07:52 crc kubenswrapper[4824]: I1209 10:07:52.107130 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" Dec 09 10:08:02 crc kubenswrapper[4824]: I1209 10:08:02.861194 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:08:02 crc kubenswrapper[4824]: I1209 10:08:02.861971 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.154285 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-r2cjc"] Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.157326 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-r2cjc" Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.168096 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.168212 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-7psdl" Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.168499 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.170824 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.174742 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-r2cjc"] Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.245620 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-lv7r4"] Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.247727 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-lv7r4" Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.250865 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.267057 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-lv7r4"] Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.329393 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59fe103c-1de9-48f4-a958-bcd0ce596060-config\") pod \"dnsmasq-dns-675f4bcbfc-r2cjc\" (UID: \"59fe103c-1de9-48f4-a958-bcd0ce596060\") " pod="openstack/dnsmasq-dns-675f4bcbfc-r2cjc" Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.329524 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwcnk\" (UniqueName: \"kubernetes.io/projected/59fe103c-1de9-48f4-a958-bcd0ce596060-kube-api-access-lwcnk\") pod \"dnsmasq-dns-675f4bcbfc-r2cjc\" (UID: \"59fe103c-1de9-48f4-a958-bcd0ce596060\") " pod="openstack/dnsmasq-dns-675f4bcbfc-r2cjc" Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.431050 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwcnk\" (UniqueName: \"kubernetes.io/projected/59fe103c-1de9-48f4-a958-bcd0ce596060-kube-api-access-lwcnk\") pod \"dnsmasq-dns-675f4bcbfc-r2cjc\" (UID: \"59fe103c-1de9-48f4-a958-bcd0ce596060\") " pod="openstack/dnsmasq-dns-675f4bcbfc-r2cjc" Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.431129 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90b0ec71-ccbc-4f8c-8004-2988bcda6655-config\") pod \"dnsmasq-dns-78dd6ddcc-lv7r4\" (UID: \"90b0ec71-ccbc-4f8c-8004-2988bcda6655\") " pod="openstack/dnsmasq-dns-78dd6ddcc-lv7r4" Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.431181 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90b0ec71-ccbc-4f8c-8004-2988bcda6655-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-lv7r4\" (UID: \"90b0ec71-ccbc-4f8c-8004-2988bcda6655\") " pod="openstack/dnsmasq-dns-78dd6ddcc-lv7r4" Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.431434 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x44mt\" (UniqueName: \"kubernetes.io/projected/90b0ec71-ccbc-4f8c-8004-2988bcda6655-kube-api-access-x44mt\") pod \"dnsmasq-dns-78dd6ddcc-lv7r4\" (UID: \"90b0ec71-ccbc-4f8c-8004-2988bcda6655\") " pod="openstack/dnsmasq-dns-78dd6ddcc-lv7r4" Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.431635 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59fe103c-1de9-48f4-a958-bcd0ce596060-config\") pod \"dnsmasq-dns-675f4bcbfc-r2cjc\" (UID: \"59fe103c-1de9-48f4-a958-bcd0ce596060\") " pod="openstack/dnsmasq-dns-675f4bcbfc-r2cjc" Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.433054 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59fe103c-1de9-48f4-a958-bcd0ce596060-config\") pod \"dnsmasq-dns-675f4bcbfc-r2cjc\" (UID: \"59fe103c-1de9-48f4-a958-bcd0ce596060\") " pod="openstack/dnsmasq-dns-675f4bcbfc-r2cjc" Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.455636 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwcnk\" (UniqueName: \"kubernetes.io/projected/59fe103c-1de9-48f4-a958-bcd0ce596060-kube-api-access-lwcnk\") pod \"dnsmasq-dns-675f4bcbfc-r2cjc\" (UID: \"59fe103c-1de9-48f4-a958-bcd0ce596060\") " pod="openstack/dnsmasq-dns-675f4bcbfc-r2cjc" Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.485275 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-r2cjc" Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.535002 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90b0ec71-ccbc-4f8c-8004-2988bcda6655-config\") pod \"dnsmasq-dns-78dd6ddcc-lv7r4\" (UID: \"90b0ec71-ccbc-4f8c-8004-2988bcda6655\") " pod="openstack/dnsmasq-dns-78dd6ddcc-lv7r4" Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.535067 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90b0ec71-ccbc-4f8c-8004-2988bcda6655-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-lv7r4\" (UID: \"90b0ec71-ccbc-4f8c-8004-2988bcda6655\") " pod="openstack/dnsmasq-dns-78dd6ddcc-lv7r4" Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.535134 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x44mt\" (UniqueName: \"kubernetes.io/projected/90b0ec71-ccbc-4f8c-8004-2988bcda6655-kube-api-access-x44mt\") pod \"dnsmasq-dns-78dd6ddcc-lv7r4\" (UID: \"90b0ec71-ccbc-4f8c-8004-2988bcda6655\") " pod="openstack/dnsmasq-dns-78dd6ddcc-lv7r4" Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.536329 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90b0ec71-ccbc-4f8c-8004-2988bcda6655-config\") pod \"dnsmasq-dns-78dd6ddcc-lv7r4\" (UID: \"90b0ec71-ccbc-4f8c-8004-2988bcda6655\") " pod="openstack/dnsmasq-dns-78dd6ddcc-lv7r4" Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.536905 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90b0ec71-ccbc-4f8c-8004-2988bcda6655-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-lv7r4\" (UID: \"90b0ec71-ccbc-4f8c-8004-2988bcda6655\") " pod="openstack/dnsmasq-dns-78dd6ddcc-lv7r4" Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.557747 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x44mt\" (UniqueName: \"kubernetes.io/projected/90b0ec71-ccbc-4f8c-8004-2988bcda6655-kube-api-access-x44mt\") pod \"dnsmasq-dns-78dd6ddcc-lv7r4\" (UID: \"90b0ec71-ccbc-4f8c-8004-2988bcda6655\") " pod="openstack/dnsmasq-dns-78dd6ddcc-lv7r4" Dec 09 10:08:11 crc kubenswrapper[4824]: I1209 10:08:11.581006 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-lv7r4" Dec 09 10:08:12 crc kubenswrapper[4824]: I1209 10:08:12.139105 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-r2cjc"] Dec 09 10:08:12 crc kubenswrapper[4824]: I1209 10:08:12.150668 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-lv7r4"] Dec 09 10:08:12 crc kubenswrapper[4824]: I1209 10:08:12.307002 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-r2cjc" event={"ID":"59fe103c-1de9-48f4-a958-bcd0ce596060","Type":"ContainerStarted","Data":"6e618550834147d803f9373985194cd65d42bf2c723702e1e86bfaace61fe6a3"} Dec 09 10:08:12 crc kubenswrapper[4824]: I1209 10:08:12.308570 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-lv7r4" event={"ID":"90b0ec71-ccbc-4f8c-8004-2988bcda6655","Type":"ContainerStarted","Data":"93fb6c59cc5898d4ff6ce7ccceef8a24f3aa1206fa6be1ca8b152bb091721c8f"} Dec 09 10:08:14 crc kubenswrapper[4824]: I1209 10:08:14.157521 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-r2cjc"] Dec 09 10:08:14 crc kubenswrapper[4824]: I1209 10:08:14.192623 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-cgdx9"] Dec 09 10:08:14 crc kubenswrapper[4824]: I1209 10:08:14.194739 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-cgdx9" Dec 09 10:08:14 crc kubenswrapper[4824]: I1209 10:08:14.209431 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-cgdx9"] Dec 09 10:08:14 crc kubenswrapper[4824]: I1209 10:08:14.397053 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n79b5\" (UniqueName: \"kubernetes.io/projected/c51bee9b-4c7f-4809-8406-488db1c6223a-kube-api-access-n79b5\") pod \"dnsmasq-dns-666b6646f7-cgdx9\" (UID: \"c51bee9b-4c7f-4809-8406-488db1c6223a\") " pod="openstack/dnsmasq-dns-666b6646f7-cgdx9" Dec 09 10:08:14 crc kubenswrapper[4824]: I1209 10:08:14.397147 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c51bee9b-4c7f-4809-8406-488db1c6223a-dns-svc\") pod \"dnsmasq-dns-666b6646f7-cgdx9\" (UID: \"c51bee9b-4c7f-4809-8406-488db1c6223a\") " pod="openstack/dnsmasq-dns-666b6646f7-cgdx9" Dec 09 10:08:14 crc kubenswrapper[4824]: I1209 10:08:14.397245 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c51bee9b-4c7f-4809-8406-488db1c6223a-config\") pod \"dnsmasq-dns-666b6646f7-cgdx9\" (UID: \"c51bee9b-4c7f-4809-8406-488db1c6223a\") " pod="openstack/dnsmasq-dns-666b6646f7-cgdx9" Dec 09 10:08:14 crc kubenswrapper[4824]: I1209 10:08:14.499323 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c51bee9b-4c7f-4809-8406-488db1c6223a-dns-svc\") pod \"dnsmasq-dns-666b6646f7-cgdx9\" (UID: \"c51bee9b-4c7f-4809-8406-488db1c6223a\") " pod="openstack/dnsmasq-dns-666b6646f7-cgdx9" Dec 09 10:08:14 crc kubenswrapper[4824]: I1209 10:08:14.499470 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c51bee9b-4c7f-4809-8406-488db1c6223a-config\") pod \"dnsmasq-dns-666b6646f7-cgdx9\" (UID: \"c51bee9b-4c7f-4809-8406-488db1c6223a\") " pod="openstack/dnsmasq-dns-666b6646f7-cgdx9" Dec 09 10:08:14 crc kubenswrapper[4824]: I1209 10:08:14.499616 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n79b5\" (UniqueName: \"kubernetes.io/projected/c51bee9b-4c7f-4809-8406-488db1c6223a-kube-api-access-n79b5\") pod \"dnsmasq-dns-666b6646f7-cgdx9\" (UID: \"c51bee9b-4c7f-4809-8406-488db1c6223a\") " pod="openstack/dnsmasq-dns-666b6646f7-cgdx9" Dec 09 10:08:14 crc kubenswrapper[4824]: I1209 10:08:14.568215 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c51bee9b-4c7f-4809-8406-488db1c6223a-dns-svc\") pod \"dnsmasq-dns-666b6646f7-cgdx9\" (UID: \"c51bee9b-4c7f-4809-8406-488db1c6223a\") " pod="openstack/dnsmasq-dns-666b6646f7-cgdx9" Dec 09 10:08:14 crc kubenswrapper[4824]: I1209 10:08:14.568328 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c51bee9b-4c7f-4809-8406-488db1c6223a-config\") pod \"dnsmasq-dns-666b6646f7-cgdx9\" (UID: \"c51bee9b-4c7f-4809-8406-488db1c6223a\") " pod="openstack/dnsmasq-dns-666b6646f7-cgdx9" Dec 09 10:08:14 crc kubenswrapper[4824]: I1209 10:08:14.597745 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n79b5\" (UniqueName: \"kubernetes.io/projected/c51bee9b-4c7f-4809-8406-488db1c6223a-kube-api-access-n79b5\") pod \"dnsmasq-dns-666b6646f7-cgdx9\" (UID: \"c51bee9b-4c7f-4809-8406-488db1c6223a\") " pod="openstack/dnsmasq-dns-666b6646f7-cgdx9" Dec 09 10:08:14 crc kubenswrapper[4824]: I1209 10:08:14.722384 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-lv7r4"] Dec 09 10:08:14 crc kubenswrapper[4824]: I1209 10:08:14.762587 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-fhkvj"] Dec 09 10:08:14 crc kubenswrapper[4824]: I1209 10:08:14.765901 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-fhkvj" Dec 09 10:08:14 crc kubenswrapper[4824]: I1209 10:08:14.783908 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-fhkvj"] Dec 09 10:08:14 crc kubenswrapper[4824]: I1209 10:08:14.806854 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-fhkvj\" (UID: \"a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae\") " pod="openstack/dnsmasq-dns-57d769cc4f-fhkvj" Dec 09 10:08:14 crc kubenswrapper[4824]: I1209 10:08:14.807036 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae-config\") pod \"dnsmasq-dns-57d769cc4f-fhkvj\" (UID: \"a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae\") " pod="openstack/dnsmasq-dns-57d769cc4f-fhkvj" Dec 09 10:08:14 crc kubenswrapper[4824]: I1209 10:08:14.807110 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n2kp\" (UniqueName: \"kubernetes.io/projected/a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae-kube-api-access-9n2kp\") pod \"dnsmasq-dns-57d769cc4f-fhkvj\" (UID: \"a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae\") " pod="openstack/dnsmasq-dns-57d769cc4f-fhkvj" Dec 09 10:08:14 crc kubenswrapper[4824]: I1209 10:08:14.822520 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-cgdx9" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.380754 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae-config\") pod \"dnsmasq-dns-57d769cc4f-fhkvj\" (UID: \"a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae\") " pod="openstack/dnsmasq-dns-57d769cc4f-fhkvj" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.380859 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n2kp\" (UniqueName: \"kubernetes.io/projected/a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae-kube-api-access-9n2kp\") pod \"dnsmasq-dns-57d769cc4f-fhkvj\" (UID: \"a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae\") " pod="openstack/dnsmasq-dns-57d769cc4f-fhkvj" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.380909 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-fhkvj\" (UID: \"a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae\") " pod="openstack/dnsmasq-dns-57d769cc4f-fhkvj" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.382566 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-fhkvj\" (UID: \"a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae\") " pod="openstack/dnsmasq-dns-57d769cc4f-fhkvj" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.383329 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae-config\") pod \"dnsmasq-dns-57d769cc4f-fhkvj\" (UID: \"a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae\") " pod="openstack/dnsmasq-dns-57d769cc4f-fhkvj" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.416454 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-1"] Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.418238 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.424423 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.425025 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.425311 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.425472 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.425637 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.425631 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-cmmw6" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.425977 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.442976 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.445009 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.445794 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n2kp\" (UniqueName: \"kubernetes.io/projected/a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae-kube-api-access-9n2kp\") pod \"dnsmasq-dns-57d769cc4f-fhkvj\" (UID: \"a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae\") " pod="openstack/dnsmasq-dns-57d769cc4f-fhkvj" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.453657 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-2"] Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.455946 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.466098 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-1"] Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.490369 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.505079 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-2"] Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.592271 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/30eb761a-7bff-43a5-adfb-40e90e17e4a6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.592327 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/30eb761a-7bff-43a5-adfb-40e90e17e4a6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.592396 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-server-conf\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.592519 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-config-data\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.592673 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.592729 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.592847 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.592896 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/30eb761a-7bff-43a5-adfb-40e90e17e4a6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.592935 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30eb761a-7bff-43a5-adfb-40e90e17e4a6-config-data\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.592953 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cee7a7d4-090b-4db0-abf1-d3f861f8e594-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.592981 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/30eb761a-7bff-43a5-adfb-40e90e17e4a6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.593004 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cee7a7d4-090b-4db0-abf1-d3f861f8e594-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.593034 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-pod-info\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.593085 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.593119 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cee7a7d4-090b-4db0-abf1-d3f861f8e594-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.593170 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cee7a7d4-090b-4db0-abf1-d3f861f8e594-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.593202 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.593237 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/30eb761a-7bff-43a5-adfb-40e90e17e4a6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.593302 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.593330 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.593389 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cee7a7d4-090b-4db0-abf1-d3f861f8e594-server-conf\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.593935 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/30eb761a-7bff-43a5-adfb-40e90e17e4a6-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.593969 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct9gv\" (UniqueName: \"kubernetes.io/projected/30eb761a-7bff-43a5-adfb-40e90e17e4a6-kube-api-access-ct9gv\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.594313 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cee7a7d4-090b-4db0-abf1-d3f861f8e594-config-data\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.594412 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5pcr\" (UniqueName: \"kubernetes.io/projected/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-kube-api-access-p5pcr\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.594476 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cee7a7d4-090b-4db0-abf1-d3f861f8e594-pod-info\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.594515 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/30eb761a-7bff-43a5-adfb-40e90e17e4a6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.594542 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cee7a7d4-090b-4db0-abf1-d3f861f8e594-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.594564 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.594592 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66mpx\" (UniqueName: \"kubernetes.io/projected/cee7a7d4-090b-4db0-abf1-d3f861f8e594-kube-api-access-66mpx\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.594612 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/30eb761a-7bff-43a5-adfb-40e90e17e4a6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.594646 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cee7a7d4-090b-4db0-abf1-d3f861f8e594-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.594908 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.696355 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cee7a7d4-090b-4db0-abf1-d3f861f8e594-pod-info\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.696486 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/30eb761a-7bff-43a5-adfb-40e90e17e4a6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.696520 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.696551 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cee7a7d4-090b-4db0-abf1-d3f861f8e594-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.696594 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/30eb761a-7bff-43a5-adfb-40e90e17e4a6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.696629 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66mpx\" (UniqueName: \"kubernetes.io/projected/cee7a7d4-090b-4db0-abf1-d3f861f8e594-kube-api-access-66mpx\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.696661 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cee7a7d4-090b-4db0-abf1-d3f861f8e594-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.696703 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.696748 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/30eb761a-7bff-43a5-adfb-40e90e17e4a6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.696920 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/30eb761a-7bff-43a5-adfb-40e90e17e4a6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.696959 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-server-conf\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.696994 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-config-data\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.697078 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.697141 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.697656 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cee7a7d4-090b-4db0-abf1-d3f861f8e594-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.698070 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cee7a7d4-090b-4db0-abf1-d3f861f8e594-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.699758 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-server-conf\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.701117 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-config-data\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.697178 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.701292 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/30eb761a-7bff-43a5-adfb-40e90e17e4a6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.701337 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30eb761a-7bff-43a5-adfb-40e90e17e4a6-config-data\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.701366 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cee7a7d4-090b-4db0-abf1-d3f861f8e594-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.701401 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/30eb761a-7bff-43a5-adfb-40e90e17e4a6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.701425 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cee7a7d4-090b-4db0-abf1-d3f861f8e594-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.701459 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-pod-info\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.701511 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.701545 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cee7a7d4-090b-4db0-abf1-d3f861f8e594-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.701587 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cee7a7d4-090b-4db0-abf1-d3f861f8e594-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.701620 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.701654 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/30eb761a-7bff-43a5-adfb-40e90e17e4a6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.701689 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.701699 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/30eb761a-7bff-43a5-adfb-40e90e17e4a6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.701722 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.701915 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cee7a7d4-090b-4db0-abf1-d3f861f8e594-server-conf\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.701974 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/30eb761a-7bff-43a5-adfb-40e90e17e4a6-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.702007 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct9gv\" (UniqueName: \"kubernetes.io/projected/30eb761a-7bff-43a5-adfb-40e90e17e4a6-kube-api-access-ct9gv\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.702069 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cee7a7d4-090b-4db0-abf1-d3f861f8e594-config-data\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.702115 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5pcr\" (UniqueName: \"kubernetes.io/projected/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-kube-api-access-p5pcr\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.702180 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/30eb761a-7bff-43a5-adfb-40e90e17e4a6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.702258 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.703002 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/30eb761a-7bff-43a5-adfb-40e90e17e4a6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.703802 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cee7a7d4-090b-4db0-abf1-d3f861f8e594-server-conf\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.703943 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.703971 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4f5f5da8e49a6f0edf98a2a4819436f9d70671dd23e36f231194d4ff6d8febbe/globalmount\"" pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.703981 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.704005 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/da6b281685d3b39c9f4d80b46538e4c0d69a86be120be6d56528f21843928b2b/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.704078 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cee7a7d4-090b-4db0-abf1-d3f861f8e594-pod-info\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.704127 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/30eb761a-7bff-43a5-adfb-40e90e17e4a6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.704309 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.705017 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.705721 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/30eb761a-7bff-43a5-adfb-40e90e17e4a6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.707062 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cee7a7d4-090b-4db0-abf1-d3f861f8e594-config-data\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.707124 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.707166 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4a25c77492aa5c7e92a555e527d30ff343f34e4f3d4bb5f823cc7b33af361462/globalmount\"" pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.707718 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cee7a7d4-090b-4db0-abf1-d3f861f8e594-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.708384 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cee7a7d4-090b-4db0-abf1-d3f861f8e594-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.709010 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30eb761a-7bff-43a5-adfb-40e90e17e4a6-config-data\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.720148 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cee7a7d4-090b-4db0-abf1-d3f861f8e594-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.720550 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/30eb761a-7bff-43a5-adfb-40e90e17e4a6-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.721168 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-pod-info\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.721479 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.722961 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-fhkvj" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.726994 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.727125 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct9gv\" (UniqueName: \"kubernetes.io/projected/30eb761a-7bff-43a5-adfb-40e90e17e4a6-kube-api-access-ct9gv\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.731731 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/30eb761a-7bff-43a5-adfb-40e90e17e4a6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.731791 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/30eb761a-7bff-43a5-adfb-40e90e17e4a6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.731978 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5pcr\" (UniqueName: \"kubernetes.io/projected/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-kube-api-access-p5pcr\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.736007 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66mpx\" (UniqueName: \"kubernetes.io/projected/cee7a7d4-090b-4db0-abf1-d3f861f8e594-kube-api-access-66mpx\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.750458 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cee7a7d4-090b-4db0-abf1-d3f861f8e594-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.775328 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.811930 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0\") pod \"rabbitmq-server-1\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.824771 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c\") pod \"rabbitmq-server-2\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " pod="openstack/rabbitmq-server-2" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.837770 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.924993 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98\") pod \"rabbitmq-server-0\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:08:15 crc kubenswrapper[4824]: I1209 10:08:15.927640 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.184974 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.260102 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-cgdx9"] Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.457284 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-cgdx9" event={"ID":"c51bee9b-4c7f-4809-8406-488db1c6223a","Type":"ContainerStarted","Data":"a167c5aec82479abbf9724d7afb2e62edd060fdf30226a87aa6fdcd62a3a3161"} Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.484194 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.486988 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.489067 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-z2n6r" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.489283 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.489511 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.489698 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.489942 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.490137 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.490342 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.516903 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.650124 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/07c1822c-1257-4d8c-9a17-d921929000fa-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.650209 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/07c1822c-1257-4d8c-9a17-d921929000fa-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.650246 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/07c1822c-1257-4d8c-9a17-d921929000fa-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.650285 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/07c1822c-1257-4d8c-9a17-d921929000fa-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.650340 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/07c1822c-1257-4d8c-9a17-d921929000fa-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.650387 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/07c1822c-1257-4d8c-9a17-d921929000fa-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.650434 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzd8c\" (UniqueName: \"kubernetes.io/projected/07c1822c-1257-4d8c-9a17-d921929000fa-kube-api-access-kzd8c\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.650467 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/07c1822c-1257-4d8c-9a17-d921929000fa-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.650525 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3785e319-25c8-406d-9ae9-497d53624c0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3785e319-25c8-406d-9ae9-497d53624c0e\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.650559 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/07c1822c-1257-4d8c-9a17-d921929000fa-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.650594 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/07c1822c-1257-4d8c-9a17-d921929000fa-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.676637 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-2"] Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.752153 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/07c1822c-1257-4d8c-9a17-d921929000fa-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.752244 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzd8c\" (UniqueName: \"kubernetes.io/projected/07c1822c-1257-4d8c-9a17-d921929000fa-kube-api-access-kzd8c\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.752304 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/07c1822c-1257-4d8c-9a17-d921929000fa-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.752357 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3785e319-25c8-406d-9ae9-497d53624c0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3785e319-25c8-406d-9ae9-497d53624c0e\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.752381 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/07c1822c-1257-4d8c-9a17-d921929000fa-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.752412 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/07c1822c-1257-4d8c-9a17-d921929000fa-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.752482 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/07c1822c-1257-4d8c-9a17-d921929000fa-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.752534 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/07c1822c-1257-4d8c-9a17-d921929000fa-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.752571 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/07c1822c-1257-4d8c-9a17-d921929000fa-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.752602 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/07c1822c-1257-4d8c-9a17-d921929000fa-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.752661 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/07c1822c-1257-4d8c-9a17-d921929000fa-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.753892 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/07c1822c-1257-4d8c-9a17-d921929000fa-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.755173 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/07c1822c-1257-4d8c-9a17-d921929000fa-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.756910 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/07c1822c-1257-4d8c-9a17-d921929000fa-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.757675 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/07c1822c-1257-4d8c-9a17-d921929000fa-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.759233 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/07c1822c-1257-4d8c-9a17-d921929000fa-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.764690 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/07c1822c-1257-4d8c-9a17-d921929000fa-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.765948 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/07c1822c-1257-4d8c-9a17-d921929000fa-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.766583 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/07c1822c-1257-4d8c-9a17-d921929000fa-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.767226 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/07c1822c-1257-4d8c-9a17-d921929000fa-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.767828 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.767856 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3785e319-25c8-406d-9ae9-497d53624c0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3785e319-25c8-406d-9ae9-497d53624c0e\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/67b607a56a6ba5190d5a998ccfc91a99cdd5ecc60a8dca58e454bce5bec86124/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.807115 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzd8c\" (UniqueName: \"kubernetes.io/projected/07c1822c-1257-4d8c-9a17-d921929000fa-kube-api-access-kzd8c\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.827402 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-fhkvj"] Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.926003 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-1"] Dec 09 10:08:16 crc kubenswrapper[4824]: I1209 10:08:16.926284 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3785e319-25c8-406d-9ae9-497d53624c0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3785e319-25c8-406d-9ae9-497d53624c0e\") pod \"rabbitmq-cell1-server-0\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.053274 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.055634 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.062322 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-xtdxd" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.063202 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.063775 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.070805 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.079653 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.090681 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.129421 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.241013 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/56148e0f-636f-410e-bfb5-342da01e8c76-kolla-config\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") " pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.241084 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/56148e0f-636f-410e-bfb5-342da01e8c76-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") " pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.241132 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56148e0f-636f-410e-bfb5-342da01e8c76-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") " pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.241219 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/56148e0f-636f-410e-bfb5-342da01e8c76-config-data-default\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") " pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.241250 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kb87\" (UniqueName: \"kubernetes.io/projected/56148e0f-636f-410e-bfb5-342da01e8c76-kube-api-access-2kb87\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") " pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.241283 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/56148e0f-636f-410e-bfb5-342da01e8c76-config-data-generated\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") " pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.241337 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b49f4990-a700-4ad8-ada0-869d524df345\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b49f4990-a700-4ad8-ada0-869d524df345\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") " pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.241378 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56148e0f-636f-410e-bfb5-342da01e8c76-operator-scripts\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") " pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.394267 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/56148e0f-636f-410e-bfb5-342da01e8c76-kolla-config\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") " pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.394341 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/56148e0f-636f-410e-bfb5-342da01e8c76-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") " pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.394393 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56148e0f-636f-410e-bfb5-342da01e8c76-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") " pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.394516 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/56148e0f-636f-410e-bfb5-342da01e8c76-config-data-default\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") " pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.394564 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kb87\" (UniqueName: \"kubernetes.io/projected/56148e0f-636f-410e-bfb5-342da01e8c76-kube-api-access-2kb87\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") " pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.394601 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/56148e0f-636f-410e-bfb5-342da01e8c76-config-data-generated\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") " pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.394672 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b49f4990-a700-4ad8-ada0-869d524df345\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b49f4990-a700-4ad8-ada0-869d524df345\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") " pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.394703 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56148e0f-636f-410e-bfb5-342da01e8c76-operator-scripts\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") " pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.396480 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56148e0f-636f-410e-bfb5-342da01e8c76-operator-scripts\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") " pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.397047 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/56148e0f-636f-410e-bfb5-342da01e8c76-kolla-config\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") " pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.405483 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/56148e0f-636f-410e-bfb5-342da01e8c76-config-data-generated\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") " pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.405550 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56148e0f-636f-410e-bfb5-342da01e8c76-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") " pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.406160 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/56148e0f-636f-410e-bfb5-342da01e8c76-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") " pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.412013 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/56148e0f-636f-410e-bfb5-342da01e8c76-config-data-default\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") " pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.425923 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.426020 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b49f4990-a700-4ad8-ada0-869d524df345\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b49f4990-a700-4ad8-ada0-869d524df345\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e64734d873ea30697570c17d3aa4aac4f5780492fc63f673b3c486696775d81f/globalmount\"" pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.429212 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kb87\" (UniqueName: \"kubernetes.io/projected/56148e0f-636f-410e-bfb5-342da01e8c76-kube-api-access-2kb87\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") " pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.445967 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.492708 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b49f4990-a700-4ad8-ada0-869d524df345\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b49f4990-a700-4ad8-ada0-869d524df345\") pod \"openstack-galera-0\" (UID: \"56148e0f-636f-410e-bfb5-342da01e8c76\") " pod="openstack/openstack-galera-0" Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.511989 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55","Type":"ContainerStarted","Data":"01b6e17da8db7c97663cfd93d2c005495e908868a882dc54077e7be6cff51a4f"} Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.517012 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"cee7a7d4-090b-4db0-abf1-d3f861f8e594","Type":"ContainerStarted","Data":"485b03fcc8ab59181a8886f62c62c507879b0e3ab6211d14df87de77b20ae512"} Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.523919 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-fhkvj" event={"ID":"a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae","Type":"ContainerStarted","Data":"2baadce7f2404f87cc1cc9f11b0fc8dbc2f4b841f052433b72ece8c232c9e4b7"} Dec 09 10:08:17 crc kubenswrapper[4824]: I1209 10:08:17.743861 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.102962 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.294214 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.563454 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"07c1822c-1257-4d8c-9a17-d921929000fa","Type":"ContainerStarted","Data":"bc57fe0d8effd04b0a062b79287d0203b9abfd60d82ab15741c68514d9dab845"} Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.599893 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.601712 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.607160 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.608734 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-rhr7x" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.608995 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.614151 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"30eb761a-7bff-43a5-adfb-40e90e17e4a6","Type":"ContainerStarted","Data":"fb119ab02c0c550da4f09eea1b5b56e3ca8d71a6d88b0573aceb4930df25f9d6"} Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.638583 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"56148e0f-636f-410e-bfb5-342da01e8c76","Type":"ContainerStarted","Data":"9ec25fd75cea2d18be58e3ad97a13f03696bee49c3d9ee289f355c8762939a58"} Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.648440 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.669950 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.672712 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.682590 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.684210 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.684343 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.684586 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-fqvlj" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.742223 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.753253 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4afe9ced-d6b6-4684-96f3-a9a419711cc7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"4afe9ced-d6b6-4684-96f3-a9a419711cc7\") " pod="openstack/memcached-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.753581 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4afe9ced-d6b6-4684-96f3-a9a419711cc7-kolla-config\") pod \"memcached-0\" (UID: \"4afe9ced-d6b6-4684-96f3-a9a419711cc7\") " pod="openstack/memcached-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.754689 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh9fm\" (UniqueName: \"kubernetes.io/projected/4afe9ced-d6b6-4684-96f3-a9a419711cc7-kube-api-access-mh9fm\") pod \"memcached-0\" (UID: \"4afe9ced-d6b6-4684-96f3-a9a419711cc7\") " pod="openstack/memcached-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.755435 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4afe9ced-d6b6-4684-96f3-a9a419711cc7-config-data\") pod \"memcached-0\" (UID: \"4afe9ced-d6b6-4684-96f3-a9a419711cc7\") " pod="openstack/memcached-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.755679 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4afe9ced-d6b6-4684-96f3-a9a419711cc7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"4afe9ced-d6b6-4684-96f3-a9a419711cc7\") " pod="openstack/memcached-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.861933 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/73d6bd70-44c7-4eed-a93a-36df636869cf-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") " pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.862254 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/73d6bd70-44c7-4eed-a93a-36df636869cf-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") " pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.862367 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4afe9ced-d6b6-4684-96f3-a9a419711cc7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"4afe9ced-d6b6-4684-96f3-a9a419711cc7\") " pod="openstack/memcached-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.862516 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4afe9ced-d6b6-4684-96f3-a9a419711cc7-kolla-config\") pod \"memcached-0\" (UID: \"4afe9ced-d6b6-4684-96f3-a9a419711cc7\") " pod="openstack/memcached-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.862628 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/73d6bd70-44c7-4eed-a93a-36df636869cf-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") " pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.862765 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/73d6bd70-44c7-4eed-a93a-36df636869cf-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") " pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.862963 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-10128d34-e475-4e75-a260-e07c3631c430\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-10128d34-e475-4e75-a260-e07c3631c430\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") " pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.863145 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73d6bd70-44c7-4eed-a93a-36df636869cf-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") " pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.863271 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh9fm\" (UniqueName: \"kubernetes.io/projected/4afe9ced-d6b6-4684-96f3-a9a419711cc7-kube-api-access-mh9fm\") pod \"memcached-0\" (UID: \"4afe9ced-d6b6-4684-96f3-a9a419711cc7\") " pod="openstack/memcached-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.863385 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqfg2\" (UniqueName: \"kubernetes.io/projected/73d6bd70-44c7-4eed-a93a-36df636869cf-kube-api-access-rqfg2\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") " pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.863521 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4afe9ced-d6b6-4684-96f3-a9a419711cc7-config-data\") pod \"memcached-0\" (UID: \"4afe9ced-d6b6-4684-96f3-a9a419711cc7\") " pod="openstack/memcached-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.863615 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4afe9ced-d6b6-4684-96f3-a9a419711cc7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"4afe9ced-d6b6-4684-96f3-a9a419711cc7\") " pod="openstack/memcached-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.863730 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73d6bd70-44c7-4eed-a93a-36df636869cf-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") " pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.867683 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4afe9ced-d6b6-4684-96f3-a9a419711cc7-kolla-config\") pod \"memcached-0\" (UID: \"4afe9ced-d6b6-4684-96f3-a9a419711cc7\") " pod="openstack/memcached-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.869219 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4afe9ced-d6b6-4684-96f3-a9a419711cc7-config-data\") pod \"memcached-0\" (UID: \"4afe9ced-d6b6-4684-96f3-a9a419711cc7\") " pod="openstack/memcached-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.873154 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4afe9ced-d6b6-4684-96f3-a9a419711cc7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"4afe9ced-d6b6-4684-96f3-a9a419711cc7\") " pod="openstack/memcached-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.891653 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh9fm\" (UniqueName: \"kubernetes.io/projected/4afe9ced-d6b6-4684-96f3-a9a419711cc7-kube-api-access-mh9fm\") pod \"memcached-0\" (UID: \"4afe9ced-d6b6-4684-96f3-a9a419711cc7\") " pod="openstack/memcached-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.897494 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4afe9ced-d6b6-4684-96f3-a9a419711cc7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"4afe9ced-d6b6-4684-96f3-a9a419711cc7\") " pod="openstack/memcached-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.953204 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.965497 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/73d6bd70-44c7-4eed-a93a-36df636869cf-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") " pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.965556 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/73d6bd70-44c7-4eed-a93a-36df636869cf-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") " pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.965684 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-10128d34-e475-4e75-a260-e07c3631c430\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-10128d34-e475-4e75-a260-e07c3631c430\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") " pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.965846 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73d6bd70-44c7-4eed-a93a-36df636869cf-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") " pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.966023 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqfg2\" (UniqueName: \"kubernetes.io/projected/73d6bd70-44c7-4eed-a93a-36df636869cf-kube-api-access-rqfg2\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") " pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.966106 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73d6bd70-44c7-4eed-a93a-36df636869cf-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") " pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.966190 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/73d6bd70-44c7-4eed-a93a-36df636869cf-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") " pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.966242 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/73d6bd70-44c7-4eed-a93a-36df636869cf-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") " pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.967851 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/73d6bd70-44c7-4eed-a93a-36df636869cf-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") " pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.968679 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/73d6bd70-44c7-4eed-a93a-36df636869cf-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") " pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.968684 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/73d6bd70-44c7-4eed-a93a-36df636869cf-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") " pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.969024 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73d6bd70-44c7-4eed-a93a-36df636869cf-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") " pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.970413 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/73d6bd70-44c7-4eed-a93a-36df636869cf-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") " pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.975586 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.975645 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-10128d34-e475-4e75-a260-e07c3631c430\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-10128d34-e475-4e75-a260-e07c3631c430\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/71d95efd239ab8bf01fe246d6f10686e514659fa05537deed29d28818b1346eb/globalmount\"" pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.977753 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73d6bd70-44c7-4eed-a93a-36df636869cf-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") " pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:18 crc kubenswrapper[4824]: I1209 10:08:18.995735 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqfg2\" (UniqueName: \"kubernetes.io/projected/73d6bd70-44c7-4eed-a93a-36df636869cf-kube-api-access-rqfg2\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") " pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:19 crc kubenswrapper[4824]: I1209 10:08:19.053526 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-10128d34-e475-4e75-a260-e07c3631c430\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-10128d34-e475-4e75-a260-e07c3631c430\") pod \"openstack-cell1-galera-0\" (UID: \"73d6bd70-44c7-4eed-a93a-36df636869cf\") " pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:19 crc kubenswrapper[4824]: I1209 10:08:19.347968 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:21 crc kubenswrapper[4824]: I1209 10:08:21.059388 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 10:08:21 crc kubenswrapper[4824]: I1209 10:08:21.061572 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 10:08:21 crc kubenswrapper[4824]: I1209 10:08:21.066008 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsvhp\" (UniqueName: \"kubernetes.io/projected/94d75a5f-1730-4993-9948-c42c98719163-kube-api-access-hsvhp\") pod \"kube-state-metrics-0\" (UID: \"94d75a5f-1730-4993-9948-c42c98719163\") " pod="openstack/kube-state-metrics-0" Dec 09 10:08:21 crc kubenswrapper[4824]: I1209 10:08:21.086268 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 10:08:21 crc kubenswrapper[4824]: I1209 10:08:21.093418 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-dlqnr" Dec 09 10:08:21 crc kubenswrapper[4824]: I1209 10:08:21.181829 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsvhp\" (UniqueName: \"kubernetes.io/projected/94d75a5f-1730-4993-9948-c42c98719163-kube-api-access-hsvhp\") pod \"kube-state-metrics-0\" (UID: \"94d75a5f-1730-4993-9948-c42c98719163\") " pod="openstack/kube-state-metrics-0" Dec 09 10:08:21 crc kubenswrapper[4824]: I1209 10:08:21.243356 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsvhp\" (UniqueName: \"kubernetes.io/projected/94d75a5f-1730-4993-9948-c42c98719163-kube-api-access-hsvhp\") pod \"kube-state-metrics-0\" (UID: \"94d75a5f-1730-4993-9948-c42c98719163\") " pod="openstack/kube-state-metrics-0" Dec 09 10:08:21 crc kubenswrapper[4824]: I1209 10:08:21.288614 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 09 10:08:21 crc kubenswrapper[4824]: I1209 10:08:21.682743 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 10:08:21 crc kubenswrapper[4824]: I1209 10:08:21.797970 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 09 10:08:21 crc kubenswrapper[4824]: I1209 10:08:21.823914 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"73d6bd70-44c7-4eed-a93a-36df636869cf","Type":"ContainerStarted","Data":"1fa32ca473dda771fcf226cf18efbbdddc2ad8aaddfc1818e389cc2c620c4cc0"} Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.533176 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.560861 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.570240 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.570300 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.570393 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.570435 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.574150 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.578205 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.578600 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-qb6sm" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.624696 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8edb90f3-f84b-4700-993f-21ffa950b620-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.624834 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8edb90f3-f84b-4700-993f-21ffa950b620-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.624855 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8edb90f3-f84b-4700-993f-21ffa950b620-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.624914 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8edb90f3-f84b-4700-993f-21ffa950b620-config\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.625087 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8edb90f3-f84b-4700-993f-21ffa950b620-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.627492 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8edb90f3-f84b-4700-993f-21ffa950b620-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.627585 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-60a80b73-7299-419c-96ae-9bc835031b04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a80b73-7299-419c-96ae-9bc835031b04\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.627657 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-628jk\" (UniqueName: \"kubernetes.io/projected/8edb90f3-f84b-4700-993f-21ffa950b620-kube-api-access-628jk\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.734431 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8edb90f3-f84b-4700-993f-21ffa950b620-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.734682 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8edb90f3-f84b-4700-993f-21ffa950b620-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.734772 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-60a80b73-7299-419c-96ae-9bc835031b04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a80b73-7299-419c-96ae-9bc835031b04\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.734878 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-628jk\" (UniqueName: \"kubernetes.io/projected/8edb90f3-f84b-4700-993f-21ffa950b620-kube-api-access-628jk\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.735011 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8edb90f3-f84b-4700-993f-21ffa950b620-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.735115 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8edb90f3-f84b-4700-993f-21ffa950b620-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.735190 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8edb90f3-f84b-4700-993f-21ffa950b620-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.735260 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8edb90f3-f84b-4700-993f-21ffa950b620-config\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.738266 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8edb90f3-f84b-4700-993f-21ffa950b620-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.754373 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.754427 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-60a80b73-7299-419c-96ae-9bc835031b04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a80b73-7299-419c-96ae-9bc835031b04\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8a9b576ec164efc98ef3fd54f8538cb024461023d7d2d915fde5b20b4af7c4e7/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.762150 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8edb90f3-f84b-4700-993f-21ffa950b620-config\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.775936 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8edb90f3-f84b-4700-993f-21ffa950b620-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.783195 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8edb90f3-f84b-4700-993f-21ffa950b620-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.786127 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-628jk\" (UniqueName: \"kubernetes.io/projected/8edb90f3-f84b-4700-993f-21ffa950b620-kube-api-access-628jk\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.789271 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8edb90f3-f84b-4700-993f-21ffa950b620-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.798699 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8edb90f3-f84b-4700-993f-21ffa950b620-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.805814 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-8j2s2"] Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.808893 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-8j2s2" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.821877 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-8j2s2"] Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.825382 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.825629 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards-sa-dockercfg-mvhpq" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.839336 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74341961-8a97-42da-9318-4c33280e45e7-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-8j2s2\" (UID: \"74341961-8a97-42da-9318-4c33280e45e7\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-8j2s2" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.839501 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrlpt\" (UniqueName: \"kubernetes.io/projected/74341961-8a97-42da-9318-4c33280e45e7-kube-api-access-mrlpt\") pod \"observability-ui-dashboards-7d5fb4cbfb-8j2s2\" (UID: \"74341961-8a97-42da-9318-4c33280e45e7\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-8j2s2" Dec 09 10:08:22 crc kubenswrapper[4824]: I1209 10:08:22.942816 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrlpt\" (UniqueName: \"kubernetes.io/projected/74341961-8a97-42da-9318-4c33280e45e7-kube-api-access-mrlpt\") pod \"observability-ui-dashboards-7d5fb4cbfb-8j2s2\" (UID: \"74341961-8a97-42da-9318-4c33280e45e7\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-8j2s2" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.045192 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74341961-8a97-42da-9318-4c33280e45e7-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-8j2s2\" (UID: \"74341961-8a97-42da-9318-4c33280e45e7\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-8j2s2" Dec 09 10:08:23 crc kubenswrapper[4824]: E1209 10:08:23.045876 4824 secret.go:188] Couldn't get secret openshift-operators/observability-ui-dashboards: secret "observability-ui-dashboards" not found Dec 09 10:08:23 crc kubenswrapper[4824]: E1209 10:08:23.046108 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/74341961-8a97-42da-9318-4c33280e45e7-serving-cert podName:74341961-8a97-42da-9318-4c33280e45e7 nodeName:}" failed. No retries permitted until 2025-12-09 10:08:23.54608811 +0000 UTC m=+1259.880592777 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/74341961-8a97-42da-9318-4c33280e45e7-serving-cert") pod "observability-ui-dashboards-7d5fb4cbfb-8j2s2" (UID: "74341961-8a97-42da-9318-4c33280e45e7") : secret "observability-ui-dashboards" not found Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.047646 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"4afe9ced-d6b6-4684-96f3-a9a419711cc7","Type":"ContainerStarted","Data":"67deb0c07d8a5968bc1fea6007802ce01762855414783475e390cf9dd44b47fd"} Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.054761 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-60a80b73-7299-419c-96ae-9bc835031b04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a80b73-7299-419c-96ae-9bc835031b04\") pod \"prometheus-metric-storage-0\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.073611 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.095358 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrlpt\" (UniqueName: \"kubernetes.io/projected/74341961-8a97-42da-9318-4c33280e45e7-kube-api-access-mrlpt\") pod \"observability-ui-dashboards-7d5fb4cbfb-8j2s2\" (UID: \"74341961-8a97-42da-9318-4c33280e45e7\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-8j2s2" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.220772 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.316895 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7c85c585b4-tgrff"] Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.318297 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.342031 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7c85c585b4-tgrff"] Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.467223 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e61278e7-d0a8-4039-909d-c2812c8a4a81-console-oauth-config\") pod \"console-7c85c585b4-tgrff\" (UID: \"e61278e7-d0a8-4039-909d-c2812c8a4a81\") " pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.467496 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhhqc\" (UniqueName: \"kubernetes.io/projected/e61278e7-d0a8-4039-909d-c2812c8a4a81-kube-api-access-dhhqc\") pod \"console-7c85c585b4-tgrff\" (UID: \"e61278e7-d0a8-4039-909d-c2812c8a4a81\") " pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.467556 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e61278e7-d0a8-4039-909d-c2812c8a4a81-console-config\") pod \"console-7c85c585b4-tgrff\" (UID: \"e61278e7-d0a8-4039-909d-c2812c8a4a81\") " pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.467619 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e61278e7-d0a8-4039-909d-c2812c8a4a81-console-serving-cert\") pod \"console-7c85c585b4-tgrff\" (UID: \"e61278e7-d0a8-4039-909d-c2812c8a4a81\") " pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.467651 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e61278e7-d0a8-4039-909d-c2812c8a4a81-service-ca\") pod \"console-7c85c585b4-tgrff\" (UID: \"e61278e7-d0a8-4039-909d-c2812c8a4a81\") " pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.467714 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e61278e7-d0a8-4039-909d-c2812c8a4a81-trusted-ca-bundle\") pod \"console-7c85c585b4-tgrff\" (UID: \"e61278e7-d0a8-4039-909d-c2812c8a4a81\") " pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.467766 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e61278e7-d0a8-4039-909d-c2812c8a4a81-oauth-serving-cert\") pod \"console-7c85c585b4-tgrff\" (UID: \"e61278e7-d0a8-4039-909d-c2812c8a4a81\") " pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.569243 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e61278e7-d0a8-4039-909d-c2812c8a4a81-console-serving-cert\") pod \"console-7c85c585b4-tgrff\" (UID: \"e61278e7-d0a8-4039-909d-c2812c8a4a81\") " pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.569301 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e61278e7-d0a8-4039-909d-c2812c8a4a81-service-ca\") pod \"console-7c85c585b4-tgrff\" (UID: \"e61278e7-d0a8-4039-909d-c2812c8a4a81\") " pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.569336 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e61278e7-d0a8-4039-909d-c2812c8a4a81-trusted-ca-bundle\") pod \"console-7c85c585b4-tgrff\" (UID: \"e61278e7-d0a8-4039-909d-c2812c8a4a81\") " pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.569359 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e61278e7-d0a8-4039-909d-c2812c8a4a81-oauth-serving-cert\") pod \"console-7c85c585b4-tgrff\" (UID: \"e61278e7-d0a8-4039-909d-c2812c8a4a81\") " pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.569460 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e61278e7-d0a8-4039-909d-c2812c8a4a81-console-oauth-config\") pod \"console-7c85c585b4-tgrff\" (UID: \"e61278e7-d0a8-4039-909d-c2812c8a4a81\") " pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.569482 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhhqc\" (UniqueName: \"kubernetes.io/projected/e61278e7-d0a8-4039-909d-c2812c8a4a81-kube-api-access-dhhqc\") pod \"console-7c85c585b4-tgrff\" (UID: \"e61278e7-d0a8-4039-909d-c2812c8a4a81\") " pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.569531 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e61278e7-d0a8-4039-909d-c2812c8a4a81-console-config\") pod \"console-7c85c585b4-tgrff\" (UID: \"e61278e7-d0a8-4039-909d-c2812c8a4a81\") " pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.569555 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74341961-8a97-42da-9318-4c33280e45e7-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-8j2s2\" (UID: \"74341961-8a97-42da-9318-4c33280e45e7\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-8j2s2" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.570909 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e61278e7-d0a8-4039-909d-c2812c8a4a81-service-ca\") pod \"console-7c85c585b4-tgrff\" (UID: \"e61278e7-d0a8-4039-909d-c2812c8a4a81\") " pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.575965 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e61278e7-d0a8-4039-909d-c2812c8a4a81-trusted-ca-bundle\") pod \"console-7c85c585b4-tgrff\" (UID: \"e61278e7-d0a8-4039-909d-c2812c8a4a81\") " pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.578007 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e61278e7-d0a8-4039-909d-c2812c8a4a81-oauth-serving-cert\") pod \"console-7c85c585b4-tgrff\" (UID: \"e61278e7-d0a8-4039-909d-c2812c8a4a81\") " pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.578428 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e61278e7-d0a8-4039-909d-c2812c8a4a81-console-serving-cert\") pod \"console-7c85c585b4-tgrff\" (UID: \"e61278e7-d0a8-4039-909d-c2812c8a4a81\") " pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.578863 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e61278e7-d0a8-4039-909d-c2812c8a4a81-console-oauth-config\") pod \"console-7c85c585b4-tgrff\" (UID: \"e61278e7-d0a8-4039-909d-c2812c8a4a81\") " pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.595648 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74341961-8a97-42da-9318-4c33280e45e7-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-8j2s2\" (UID: \"74341961-8a97-42da-9318-4c33280e45e7\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-8j2s2" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.596704 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e61278e7-d0a8-4039-909d-c2812c8a4a81-console-config\") pod \"console-7c85c585b4-tgrff\" (UID: \"e61278e7-d0a8-4039-909d-c2812c8a4a81\") " pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.621586 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhhqc\" (UniqueName: \"kubernetes.io/projected/e61278e7-d0a8-4039-909d-c2812c8a4a81-kube-api-access-dhhqc\") pod \"console-7c85c585b4-tgrff\" (UID: \"e61278e7-d0a8-4039-909d-c2812c8a4a81\") " pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.661937 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-8j2s2" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.716409 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.717112 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.718768 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.724297 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-hw92k" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.724476 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.724888 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.725001 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.725572 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.808960 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.881150 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c22f05c5-6c0e-4e80-afff-b2642d4a7655-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") " pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.881211 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c22f05c5-6c0e-4e80-afff-b2642d4a7655-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") " pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.881298 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d9d22108-d63f-4b27-b199-e21647786f0c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9d22108-d63f-4b27-b199-e21647786f0c\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") " pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.881388 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c22f05c5-6c0e-4e80-afff-b2642d4a7655-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") " pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.881427 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c22f05c5-6c0e-4e80-afff-b2642d4a7655-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") " pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.881458 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69wdc\" (UniqueName: \"kubernetes.io/projected/c22f05c5-6c0e-4e80-afff-b2642d4a7655-kube-api-access-69wdc\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") " pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.881497 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c22f05c5-6c0e-4e80-afff-b2642d4a7655-config\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") " pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:23 crc kubenswrapper[4824]: I1209 10:08:23.881545 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c22f05c5-6c0e-4e80-afff-b2642d4a7655-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") " pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.031552 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d9d22108-d63f-4b27-b199-e21647786f0c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9d22108-d63f-4b27-b199-e21647786f0c\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") " pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.031660 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c22f05c5-6c0e-4e80-afff-b2642d4a7655-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") " pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.031704 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c22f05c5-6c0e-4e80-afff-b2642d4a7655-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") " pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.031744 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69wdc\" (UniqueName: \"kubernetes.io/projected/c22f05c5-6c0e-4e80-afff-b2642d4a7655-kube-api-access-69wdc\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") " pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.031798 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c22f05c5-6c0e-4e80-afff-b2642d4a7655-config\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") " pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.031858 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c22f05c5-6c0e-4e80-afff-b2642d4a7655-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") " pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.032040 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c22f05c5-6c0e-4e80-afff-b2642d4a7655-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") " pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.032079 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c22f05c5-6c0e-4e80-afff-b2642d4a7655-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") " pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.034894 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c22f05c5-6c0e-4e80-afff-b2642d4a7655-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") " pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.038174 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.038415 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.040056 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.041313 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.051497 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c22f05c5-6c0e-4e80-afff-b2642d4a7655-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") " pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.054623 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c22f05c5-6c0e-4e80-afff-b2642d4a7655-config\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") " pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.064409 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c22f05c5-6c0e-4e80-afff-b2642d4a7655-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") " pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.066569 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.066709 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d9d22108-d63f-4b27-b199-e21647786f0c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9d22108-d63f-4b27-b199-e21647786f0c\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/7a1e743853c537ccc2ba5861748e63c1b2585fb659e2f7b18d0a9a1896b73066/globalmount\"" pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.076446 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69wdc\" (UniqueName: \"kubernetes.io/projected/c22f05c5-6c0e-4e80-afff-b2642d4a7655-kube-api-access-69wdc\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") " pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.241530 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c22f05c5-6c0e-4e80-afff-b2642d4a7655-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") " pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.273946 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c22f05c5-6c0e-4e80-afff-b2642d4a7655-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") " pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.274294 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"94d75a5f-1730-4993-9948-c42c98719163","Type":"ContainerStarted","Data":"f61ec0a06ab6a4d4661310bad885a53a690ad0a6a4b424db05c42e6e27e1a55b"} Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.344739 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d9d22108-d63f-4b27-b199-e21647786f0c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9d22108-d63f-4b27-b199-e21647786f0c\") pod \"ovsdbserver-nb-0\" (UID: \"c22f05c5-6c0e-4e80-afff-b2642d4a7655\") " pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.344883 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.413493 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-hw92k" Dec 09 10:08:24 crc kubenswrapper[4824]: I1209 10:08:24.420848 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.051314 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-985bt"] Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.073711 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-rl4nw"] Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.076130 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-rl4nw" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.077140 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-985bt" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.077373 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-985bt"] Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.080847 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-4nsxn" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.081015 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.081394 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.090923 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-rl4nw"] Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.264644 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2a54d12b-327d-409e-8652-9525878ae96f-etc-ovs\") pod \"ovn-controller-ovs-rl4nw\" (UID: \"2a54d12b-327d-409e-8652-9525878ae96f\") " pod="openstack/ovn-controller-ovs-rl4nw" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.265060 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2a54d12b-327d-409e-8652-9525878ae96f-var-log\") pod \"ovn-controller-ovs-rl4nw\" (UID: \"2a54d12b-327d-409e-8652-9525878ae96f\") " pod="openstack/ovn-controller-ovs-rl4nw" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.265105 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2a54d12b-327d-409e-8652-9525878ae96f-var-lib\") pod \"ovn-controller-ovs-rl4nw\" (UID: \"2a54d12b-327d-409e-8652-9525878ae96f\") " pod="openstack/ovn-controller-ovs-rl4nw" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.265145 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d-var-run\") pod \"ovn-controller-985bt\" (UID: \"961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d\") " pod="openstack/ovn-controller-985bt" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.265170 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d-var-run-ovn\") pod \"ovn-controller-985bt\" (UID: \"961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d\") " pod="openstack/ovn-controller-985bt" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.265198 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpqgg\" (UniqueName: \"kubernetes.io/projected/961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d-kube-api-access-qpqgg\") pod \"ovn-controller-985bt\" (UID: \"961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d\") " pod="openstack/ovn-controller-985bt" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.265225 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2a54d12b-327d-409e-8652-9525878ae96f-scripts\") pod \"ovn-controller-ovs-rl4nw\" (UID: \"2a54d12b-327d-409e-8652-9525878ae96f\") " pod="openstack/ovn-controller-ovs-rl4nw" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.265265 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d-ovn-controller-tls-certs\") pod \"ovn-controller-985bt\" (UID: \"961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d\") " pod="openstack/ovn-controller-985bt" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.265307 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d-scripts\") pod \"ovn-controller-985bt\" (UID: \"961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d\") " pod="openstack/ovn-controller-985bt" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.265382 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d-var-log-ovn\") pod \"ovn-controller-985bt\" (UID: \"961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d\") " pod="openstack/ovn-controller-985bt" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.265411 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d-combined-ca-bundle\") pod \"ovn-controller-985bt\" (UID: \"961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d\") " pod="openstack/ovn-controller-985bt" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.265450 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9297z\" (UniqueName: \"kubernetes.io/projected/2a54d12b-327d-409e-8652-9525878ae96f-kube-api-access-9297z\") pod \"ovn-controller-ovs-rl4nw\" (UID: \"2a54d12b-327d-409e-8652-9525878ae96f\") " pod="openstack/ovn-controller-ovs-rl4nw" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.265497 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2a54d12b-327d-409e-8652-9525878ae96f-var-run\") pod \"ovn-controller-ovs-rl4nw\" (UID: \"2a54d12b-327d-409e-8652-9525878ae96f\") " pod="openstack/ovn-controller-ovs-rl4nw" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.367599 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2a54d12b-327d-409e-8652-9525878ae96f-etc-ovs\") pod \"ovn-controller-ovs-rl4nw\" (UID: \"2a54d12b-327d-409e-8652-9525878ae96f\") " pod="openstack/ovn-controller-ovs-rl4nw" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.367694 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2a54d12b-327d-409e-8652-9525878ae96f-var-log\") pod \"ovn-controller-ovs-rl4nw\" (UID: \"2a54d12b-327d-409e-8652-9525878ae96f\") " pod="openstack/ovn-controller-ovs-rl4nw" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.367729 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2a54d12b-327d-409e-8652-9525878ae96f-var-lib\") pod \"ovn-controller-ovs-rl4nw\" (UID: \"2a54d12b-327d-409e-8652-9525878ae96f\") " pod="openstack/ovn-controller-ovs-rl4nw" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.367766 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d-var-run\") pod \"ovn-controller-985bt\" (UID: \"961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d\") " pod="openstack/ovn-controller-985bt" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.367807 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d-var-run-ovn\") pod \"ovn-controller-985bt\" (UID: \"961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d\") " pod="openstack/ovn-controller-985bt" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.367833 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpqgg\" (UniqueName: \"kubernetes.io/projected/961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d-kube-api-access-qpqgg\") pod \"ovn-controller-985bt\" (UID: \"961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d\") " pod="openstack/ovn-controller-985bt" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.367863 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2a54d12b-327d-409e-8652-9525878ae96f-scripts\") pod \"ovn-controller-ovs-rl4nw\" (UID: \"2a54d12b-327d-409e-8652-9525878ae96f\") " pod="openstack/ovn-controller-ovs-rl4nw" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.367906 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d-ovn-controller-tls-certs\") pod \"ovn-controller-985bt\" (UID: \"961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d\") " pod="openstack/ovn-controller-985bt" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.367956 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d-scripts\") pod \"ovn-controller-985bt\" (UID: \"961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d\") " pod="openstack/ovn-controller-985bt" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.368040 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d-combined-ca-bundle\") pod \"ovn-controller-985bt\" (UID: \"961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d\") " pod="openstack/ovn-controller-985bt" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.368077 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d-var-log-ovn\") pod \"ovn-controller-985bt\" (UID: \"961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d\") " pod="openstack/ovn-controller-985bt" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.368144 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9297z\" (UniqueName: \"kubernetes.io/projected/2a54d12b-327d-409e-8652-9525878ae96f-kube-api-access-9297z\") pod \"ovn-controller-ovs-rl4nw\" (UID: \"2a54d12b-327d-409e-8652-9525878ae96f\") " pod="openstack/ovn-controller-ovs-rl4nw" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.368379 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2a54d12b-327d-409e-8652-9525878ae96f-var-log\") pod \"ovn-controller-ovs-rl4nw\" (UID: \"2a54d12b-327d-409e-8652-9525878ae96f\") " pod="openstack/ovn-controller-ovs-rl4nw" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.368494 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2a54d12b-327d-409e-8652-9525878ae96f-etc-ovs\") pod \"ovn-controller-ovs-rl4nw\" (UID: \"2a54d12b-327d-409e-8652-9525878ae96f\") " pod="openstack/ovn-controller-ovs-rl4nw" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.368564 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d-var-run-ovn\") pod \"ovn-controller-985bt\" (UID: \"961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d\") " pod="openstack/ovn-controller-985bt" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.368651 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d-var-run\") pod \"ovn-controller-985bt\" (UID: \"961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d\") " pod="openstack/ovn-controller-985bt" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.368653 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d-var-log-ovn\") pod \"ovn-controller-985bt\" (UID: \"961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d\") " pod="openstack/ovn-controller-985bt" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.371250 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2a54d12b-327d-409e-8652-9525878ae96f-var-lib\") pod \"ovn-controller-ovs-rl4nw\" (UID: \"2a54d12b-327d-409e-8652-9525878ae96f\") " pod="openstack/ovn-controller-ovs-rl4nw" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.372905 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2a54d12b-327d-409e-8652-9525878ae96f-var-run\") pod \"ovn-controller-ovs-rl4nw\" (UID: \"2a54d12b-327d-409e-8652-9525878ae96f\") " pod="openstack/ovn-controller-ovs-rl4nw" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.374476 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2a54d12b-327d-409e-8652-9525878ae96f-var-run\") pod \"ovn-controller-ovs-rl4nw\" (UID: \"2a54d12b-327d-409e-8652-9525878ae96f\") " pod="openstack/ovn-controller-ovs-rl4nw" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.374927 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2a54d12b-327d-409e-8652-9525878ae96f-scripts\") pod \"ovn-controller-ovs-rl4nw\" (UID: \"2a54d12b-327d-409e-8652-9525878ae96f\") " pod="openstack/ovn-controller-ovs-rl4nw" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.377437 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d-scripts\") pod \"ovn-controller-985bt\" (UID: \"961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d\") " pod="openstack/ovn-controller-985bt" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.394516 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d-ovn-controller-tls-certs\") pod \"ovn-controller-985bt\" (UID: \"961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d\") " pod="openstack/ovn-controller-985bt" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.395111 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d-combined-ca-bundle\") pod \"ovn-controller-985bt\" (UID: \"961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d\") " pod="openstack/ovn-controller-985bt" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.395548 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9297z\" (UniqueName: \"kubernetes.io/projected/2a54d12b-327d-409e-8652-9525878ae96f-kube-api-access-9297z\") pod \"ovn-controller-ovs-rl4nw\" (UID: \"2a54d12b-327d-409e-8652-9525878ae96f\") " pod="openstack/ovn-controller-ovs-rl4nw" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.409508 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpqgg\" (UniqueName: \"kubernetes.io/projected/961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d-kube-api-access-qpqgg\") pod \"ovn-controller-985bt\" (UID: \"961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d\") " pod="openstack/ovn-controller-985bt" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.443409 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-rl4nw" Dec 09 10:08:25 crc kubenswrapper[4824]: I1209 10:08:25.466686 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-985bt" Dec 09 10:08:28 crc kubenswrapper[4824]: I1209 10:08:28.989826 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 09 10:08:28 crc kubenswrapper[4824]: I1209 10:08:28.992929 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.000817 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-4ctqx" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.000999 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.001323 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.001532 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.020414 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.126981 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec252d15-4814-478d-ad29-388842d50a32-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") " pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.127052 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec252d15-4814-478d-ad29-388842d50a32-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") " pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.127163 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec252d15-4814-478d-ad29-388842d50a32-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") " pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.127211 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ec252d15-4814-478d-ad29-388842d50a32-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") " pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.127242 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpgm4\" (UniqueName: \"kubernetes.io/projected/ec252d15-4814-478d-ad29-388842d50a32-kube-api-access-kpgm4\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") " pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.127340 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec252d15-4814-478d-ad29-388842d50a32-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") " pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.127429 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec252d15-4814-478d-ad29-388842d50a32-config\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") " pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.127539 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7c334af5-a725-42d4-b262-02cd229568b6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c334af5-a725-42d4-b262-02cd229568b6\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") " pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.229528 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec252d15-4814-478d-ad29-388842d50a32-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") " pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.229582 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ec252d15-4814-478d-ad29-388842d50a32-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") " pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.229620 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpgm4\" (UniqueName: \"kubernetes.io/projected/ec252d15-4814-478d-ad29-388842d50a32-kube-api-access-kpgm4\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") " pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.229698 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec252d15-4814-478d-ad29-388842d50a32-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") " pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.229884 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec252d15-4814-478d-ad29-388842d50a32-config\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") " pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.229986 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7c334af5-a725-42d4-b262-02cd229568b6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c334af5-a725-42d4-b262-02cd229568b6\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") " pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.230083 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec252d15-4814-478d-ad29-388842d50a32-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") " pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.230139 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec252d15-4814-478d-ad29-388842d50a32-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") " pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.233138 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ec252d15-4814-478d-ad29-388842d50a32-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") " pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.234115 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec252d15-4814-478d-ad29-388842d50a32-config\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") " pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.234595 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.234625 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7c334af5-a725-42d4-b262-02cd229568b6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c334af5-a725-42d4-b262-02cd229568b6\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/18e7571243f54df72aa71cd5231c095912f529adc235c5607d09805c581c5e2a/globalmount\"" pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.235238 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec252d15-4814-478d-ad29-388842d50a32-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") " pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.238723 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec252d15-4814-478d-ad29-388842d50a32-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") " pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.240830 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec252d15-4814-478d-ad29-388842d50a32-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") " pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.247113 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec252d15-4814-478d-ad29-388842d50a32-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") " pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.249831 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpgm4\" (UniqueName: \"kubernetes.io/projected/ec252d15-4814-478d-ad29-388842d50a32-kube-api-access-kpgm4\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") " pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.293647 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7c334af5-a725-42d4-b262-02cd229568b6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c334af5-a725-42d4-b262-02cd229568b6\") pod \"ovsdbserver-sb-0\" (UID: \"ec252d15-4814-478d-ad29-388842d50a32\") " pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:29 crc kubenswrapper[4824]: I1209 10:08:29.350058 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 09 10:08:32 crc kubenswrapper[4824]: I1209 10:08:32.861393 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:08:32 crc kubenswrapper[4824]: I1209 10:08:32.862052 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:08:32 crc kubenswrapper[4824]: I1209 10:08:32.862103 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 10:08:32 crc kubenswrapper[4824]: I1209 10:08:32.862973 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b01bf8ba5eb26e2cf11005bb5cc48bd36d1aa75d59e84d3c71f092c10deef9e7"} pod="openshift-machine-config-operator/machine-config-daemon-dth8x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 10:08:32 crc kubenswrapper[4824]: I1209 10:08:32.863139 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" containerID="cri-o://b01bf8ba5eb26e2cf11005bb5cc48bd36d1aa75d59e84d3c71f092c10deef9e7" gracePeriod=600 Dec 09 10:08:33 crc kubenswrapper[4824]: E1209 10:08:33.161022 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee3c86bf_f246_42cb_adb6_900e3d735166.slice/crio-b01bf8ba5eb26e2cf11005bb5cc48bd36d1aa75d59e84d3c71f092c10deef9e7.scope\": RecentStats: unable to find data in memory cache]" Dec 09 10:08:35 crc kubenswrapper[4824]: I1209 10:08:35.638184 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerID="b01bf8ba5eb26e2cf11005bb5cc48bd36d1aa75d59e84d3c71f092c10deef9e7" exitCode=0 Dec 09 10:08:35 crc kubenswrapper[4824]: I1209 10:08:35.638290 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerDied","Data":"b01bf8ba5eb26e2cf11005bb5cc48bd36d1aa75d59e84d3c71f092c10deef9e7"} Dec 09 10:08:35 crc kubenswrapper[4824]: I1209 10:08:35.639086 4824 scope.go:117] "RemoveContainer" containerID="5abd964d470ad0e2def92ff94c5fb5265d13f56cfeabd66524120ead526f4c21" Dec 09 10:08:35 crc kubenswrapper[4824]: I1209 10:08:35.640849 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8edb90f3-f84b-4700-993f-21ffa950b620","Type":"ContainerStarted","Data":"f3654a808c4c98871b6deba78bc55172067e7f26470defbce467d95bb738e7c3"} Dec 09 10:08:35 crc kubenswrapper[4824]: I1209 10:08:35.960503 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-8j2s2"] Dec 09 10:08:36 crc kubenswrapper[4824]: W1209 10:08:36.312819 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74341961_8a97_42da_9318_4c33280e45e7.slice/crio-bfa699102ff5f8af8fd4ed0f5491e4a35e6d7445d7c6c0f187a04536cc8c3c83 WatchSource:0}: Error finding container bfa699102ff5f8af8fd4ed0f5491e4a35e6d7445d7c6c0f187a04536cc8c3c83: Status 404 returned error can't find the container with id bfa699102ff5f8af8fd4ed0f5491e4a35e6d7445d7c6c0f187a04536cc8c3c83 Dec 09 10:08:36 crc kubenswrapper[4824]: I1209 10:08:36.651502 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-8j2s2" event={"ID":"74341961-8a97-42da-9318-4c33280e45e7","Type":"ContainerStarted","Data":"bfa699102ff5f8af8fd4ed0f5491e4a35e6d7445d7c6c0f187a04536cc8c3c83"} Dec 09 10:08:36 crc kubenswrapper[4824]: I1209 10:08:36.792587 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7c85c585b4-tgrff"] Dec 09 10:08:44 crc kubenswrapper[4824]: E1209 10:08:44.115702 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 09 10:08:44 crc kubenswrapper[4824]: E1209 10:08:44.116570 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kzd8c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(07c1822c-1257-4d8c-9a17-d921929000fa): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:08:44 crc kubenswrapper[4824]: E1209 10:08:44.118120 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="07c1822c-1257-4d8c-9a17-d921929000fa" Dec 09 10:08:44 crc kubenswrapper[4824]: I1209 10:08:44.910550 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7c85c585b4-tgrff" event={"ID":"e61278e7-d0a8-4039-909d-c2812c8a4a81","Type":"ContainerStarted","Data":"1834db21c9eabaf4178e8257d23480a9aacdc741bde35a8d909ce74f93099482"} Dec 09 10:08:44 crc kubenswrapper[4824]: E1209 10:08:44.913002 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="07c1822c-1257-4d8c-9a17-d921929000fa" Dec 09 10:08:46 crc kubenswrapper[4824]: E1209 10:08:46.013768 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Dec 09 10:08:46 crc kubenswrapper[4824]: E1209 10:08:46.013970 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2kb87,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(56148e0f-636f-410e-bfb5-342da01e8c76): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:08:46 crc kubenswrapper[4824]: E1209 10:08:46.016046 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="56148e0f-636f-410e-bfb5-342da01e8c76" Dec 09 10:08:46 crc kubenswrapper[4824]: E1209 10:08:46.929648 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="56148e0f-636f-410e-bfb5-342da01e8c76" Dec 09 10:08:50 crc kubenswrapper[4824]: E1209 10:08:50.395457 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 09 10:08:50 crc kubenswrapper[4824]: E1209 10:08:50.396298 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ct9gv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(30eb761a-7bff-43a5-adfb-40e90e17e4a6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:08:50 crc kubenswrapper[4824]: E1209 10:08:50.397504 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="30eb761a-7bff-43a5-adfb-40e90e17e4a6" Dec 09 10:08:50 crc kubenswrapper[4824]: E1209 10:08:50.439688 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 09 10:08:50 crc kubenswrapper[4824]: E1209 10:08:50.439925 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-66mpx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-1_openstack(cee7a7d4-090b-4db0-abf1-d3f861f8e594): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:08:50 crc kubenswrapper[4824]: E1209 10:08:50.441138 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-1" podUID="cee7a7d4-090b-4db0-abf1-d3f861f8e594" Dec 09 10:08:50 crc kubenswrapper[4824]: E1209 10:08:50.485883 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 09 10:08:50 crc kubenswrapper[4824]: E1209 10:08:50.486077 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p5pcr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-2_openstack(09e7de63-6c82-4cb9-b9b6-12da1f3e6b55): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:08:50 crc kubenswrapper[4824]: E1209 10:08:50.487372 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-2" podUID="09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" Dec 09 10:08:50 crc kubenswrapper[4824]: E1209 10:08:50.972123 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="30eb761a-7bff-43a5-adfb-40e90e17e4a6" Dec 09 10:08:50 crc kubenswrapper[4824]: E1209 10:08:50.972301 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-2" podUID="09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" Dec 09 10:08:50 crc kubenswrapper[4824]: E1209 10:08:50.972464 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-1" podUID="cee7a7d4-090b-4db0-abf1-d3f861f8e594" Dec 09 10:08:51 crc kubenswrapper[4824]: E1209 10:08:51.154536 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached:current-podified" Dec 09 10:08:51 crc kubenswrapper[4824]: E1209 10:08:51.154736 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n89h8h64bh65h689h5bbh649h9bh6hcbh79h89h68dh76h587hcbh7fh578h5cbh5dfh57bh66ch598h55h598h5dfh564h665h5c6h549h698h75q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mh9fm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(4afe9ced-d6b6-4684-96f3-a9a419711cc7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:08:51 crc kubenswrapper[4824]: E1209 10:08:51.156751 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="4afe9ced-d6b6-4684-96f3-a9a419711cc7" Dec 09 10:08:51 crc kubenswrapper[4824]: E1209 10:08:51.966612 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 09 10:08:51 crc kubenswrapper[4824]: E1209 10:08:51.966805 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x44mt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-lv7r4_openstack(90b0ec71-ccbc-4f8c-8004-2988bcda6655): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:08:51 crc kubenswrapper[4824]: E1209 10:08:51.967897 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-lv7r4" podUID="90b0ec71-ccbc-4f8c-8004-2988bcda6655" Dec 09 10:08:51 crc kubenswrapper[4824]: E1209 10:08:51.969860 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 09 10:08:51 crc kubenswrapper[4824]: E1209 10:08:51.970041 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9n2kp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-fhkvj_openstack(a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:08:51 crc kubenswrapper[4824]: E1209 10:08:51.971847 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-fhkvj" podUID="a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae" Dec 09 10:08:51 crc kubenswrapper[4824]: E1209 10:08:51.984699 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached:current-podified\\\"\"" pod="openstack/memcached-0" podUID="4afe9ced-d6b6-4684-96f3-a9a419711cc7" Dec 09 10:08:51 crc kubenswrapper[4824]: E1209 10:08:51.988367 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 09 10:08:51 crc kubenswrapper[4824]: E1209 10:08:51.988734 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lwcnk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-r2cjc_openstack(59fe103c-1de9-48f4-a958-bcd0ce596060): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:08:51 crc kubenswrapper[4824]: E1209 10:08:51.991149 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-r2cjc" podUID="59fe103c-1de9-48f4-a958-bcd0ce596060" Dec 09 10:08:52 crc kubenswrapper[4824]: E1209 10:08:52.469300 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:1133c973c7472c665f910a722e19c8e2e27accb34b90fab67f14548627ce9c62" Dec 09 10:08:52 crc kubenswrapper[4824]: E1209 10:08:52.469768 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init-config-reloader,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:1133c973c7472c665f910a722e19c8e2e27accb34b90fab67f14548627ce9c62,Command:[/bin/prometheus-config-reloader],Args:[--watch-interval=0 --listen-address=:8081 --config-file=/etc/prometheus/config/prometheus.yaml.gz --config-envsubst-file=/etc/prometheus/config_out/prometheus.env.yaml --watched-dir=/etc/prometheus/rules/prometheus-metric-storage-rulefiles-0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:reloader-init,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:SHARD,Value:0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:false,MountPath:/etc/prometheus/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-out,ReadOnly:false,MountPath:/etc/prometheus/config_out,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:prometheus-metric-storage-rulefiles-0,ReadOnly:false,MountPath:/etc/prometheus/rules/prometheus-metric-storage-rulefiles-0,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-628jk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod prometheus-metric-storage-0_openstack(8edb90f3-f84b-4700-993f-21ffa950b620): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 10:08:52 crc kubenswrapper[4824]: E1209 10:08:52.471834 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init-config-reloader\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/prometheus-metric-storage-0" podUID="8edb90f3-f84b-4700-993f-21ffa950b620" Dec 09 10:08:52 crc kubenswrapper[4824]: E1209 10:08:52.583221 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 09 10:08:52 crc kubenswrapper[4824]: E1209 10:08:52.583683 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n79b5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-cgdx9_openstack(c51bee9b-4c7f-4809-8406-488db1c6223a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:08:52 crc kubenswrapper[4824]: E1209 10:08:52.585085 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-cgdx9" podUID="c51bee9b-4c7f-4809-8406-488db1c6223a" Dec 09 10:08:52 crc kubenswrapper[4824]: E1209 10:08:52.995769 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-fhkvj" podUID="a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae" Dec 09 10:08:52 crc kubenswrapper[4824]: E1209 10:08:52.997950 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init-config-reloader\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:1133c973c7472c665f910a722e19c8e2e27accb34b90fab67f14548627ce9c62\\\"\"" pod="openstack/prometheus-metric-storage-0" podUID="8edb90f3-f84b-4700-993f-21ffa950b620" Dec 09 10:08:52 crc kubenswrapper[4824]: E1209 10:08:52.998234 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-cgdx9" podUID="c51bee9b-4c7f-4809-8406-488db1c6223a" Dec 09 10:08:53 crc kubenswrapper[4824]: I1209 10:08:53.080195 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-985bt"] Dec 09 10:08:53 crc kubenswrapper[4824]: I1209 10:08:53.097906 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 09 10:08:53 crc kubenswrapper[4824]: I1209 10:08:53.179096 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 09 10:08:53 crc kubenswrapper[4824]: I1209 10:08:53.326342 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-rl4nw"] Dec 09 10:08:53 crc kubenswrapper[4824]: I1209 10:08:53.427215 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-lv7r4" Dec 09 10:08:53 crc kubenswrapper[4824]: W1209 10:08:53.483041 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec252d15_4814_478d_ad29_388842d50a32.slice/crio-ecbb9c2a8fcecabe62ed91b79a214788b05a2f302d5cec36bfafb4429fd5207a WatchSource:0}: Error finding container ecbb9c2a8fcecabe62ed91b79a214788b05a2f302d5cec36bfafb4429fd5207a: Status 404 returned error can't find the container with id ecbb9c2a8fcecabe62ed91b79a214788b05a2f302d5cec36bfafb4429fd5207a Dec 09 10:08:53 crc kubenswrapper[4824]: W1209 10:08:53.484229 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod961b1a8a_2fa7_4b53_bf91_c28a6d2fea6d.slice/crio-03cd7f58b1f6c7fc894a911249c750f8c1a2b4e4b4382cd5802d849d81143d26 WatchSource:0}: Error finding container 03cd7f58b1f6c7fc894a911249c750f8c1a2b4e4b4382cd5802d849d81143d26: Status 404 returned error can't find the container with id 03cd7f58b1f6c7fc894a911249c750f8c1a2b4e4b4382cd5802d849d81143d26 Dec 09 10:08:53 crc kubenswrapper[4824]: W1209 10:08:53.492061 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a54d12b_327d_409e_8652_9525878ae96f.slice/crio-0d78048d5100ff49d4ee84d6408a228431dcb634fa5fa74e08dcfeb6453fa1ce WatchSource:0}: Error finding container 0d78048d5100ff49d4ee84d6408a228431dcb634fa5fa74e08dcfeb6453fa1ce: Status 404 returned error can't find the container with id 0d78048d5100ff49d4ee84d6408a228431dcb634fa5fa74e08dcfeb6453fa1ce Dec 09 10:08:53 crc kubenswrapper[4824]: I1209 10:08:53.497869 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x44mt\" (UniqueName: \"kubernetes.io/projected/90b0ec71-ccbc-4f8c-8004-2988bcda6655-kube-api-access-x44mt\") pod \"90b0ec71-ccbc-4f8c-8004-2988bcda6655\" (UID: \"90b0ec71-ccbc-4f8c-8004-2988bcda6655\") " Dec 09 10:08:53 crc kubenswrapper[4824]: I1209 10:08:53.498921 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90b0ec71-ccbc-4f8c-8004-2988bcda6655-dns-svc\") pod \"90b0ec71-ccbc-4f8c-8004-2988bcda6655\" (UID: \"90b0ec71-ccbc-4f8c-8004-2988bcda6655\") " Dec 09 10:08:53 crc kubenswrapper[4824]: I1209 10:08:53.498962 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90b0ec71-ccbc-4f8c-8004-2988bcda6655-config\") pod \"90b0ec71-ccbc-4f8c-8004-2988bcda6655\" (UID: \"90b0ec71-ccbc-4f8c-8004-2988bcda6655\") " Dec 09 10:08:53 crc kubenswrapper[4824]: I1209 10:08:53.499699 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90b0ec71-ccbc-4f8c-8004-2988bcda6655-config" (OuterVolumeSpecName: "config") pod "90b0ec71-ccbc-4f8c-8004-2988bcda6655" (UID: "90b0ec71-ccbc-4f8c-8004-2988bcda6655"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:08:53 crc kubenswrapper[4824]: W1209 10:08:53.499973 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc22f05c5_6c0e_4e80_afff_b2642d4a7655.slice/crio-0dac3e64c1bb4d1fe419d6750fa4b8d78e24e43689628fc4875b72aad46cff92 WatchSource:0}: Error finding container 0dac3e64c1bb4d1fe419d6750fa4b8d78e24e43689628fc4875b72aad46cff92: Status 404 returned error can't find the container with id 0dac3e64c1bb4d1fe419d6750fa4b8d78e24e43689628fc4875b72aad46cff92 Dec 09 10:08:53 crc kubenswrapper[4824]: I1209 10:08:53.500356 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90b0ec71-ccbc-4f8c-8004-2988bcda6655-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "90b0ec71-ccbc-4f8c-8004-2988bcda6655" (UID: "90b0ec71-ccbc-4f8c-8004-2988bcda6655"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:08:53 crc kubenswrapper[4824]: I1209 10:08:53.507920 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90b0ec71-ccbc-4f8c-8004-2988bcda6655-kube-api-access-x44mt" (OuterVolumeSpecName: "kube-api-access-x44mt") pod "90b0ec71-ccbc-4f8c-8004-2988bcda6655" (UID: "90b0ec71-ccbc-4f8c-8004-2988bcda6655"). InnerVolumeSpecName "kube-api-access-x44mt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:08:53 crc kubenswrapper[4824]: I1209 10:08:53.583205 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-r2cjc" Dec 09 10:08:53 crc kubenswrapper[4824]: I1209 10:08:53.603077 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90b0ec71-ccbc-4f8c-8004-2988bcda6655-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 10:08:53 crc kubenswrapper[4824]: I1209 10:08:53.603113 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90b0ec71-ccbc-4f8c-8004-2988bcda6655-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:08:53 crc kubenswrapper[4824]: I1209 10:08:53.603123 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x44mt\" (UniqueName: \"kubernetes.io/projected/90b0ec71-ccbc-4f8c-8004-2988bcda6655-kube-api-access-x44mt\") on node \"crc\" DevicePath \"\"" Dec 09 10:08:53 crc kubenswrapper[4824]: I1209 10:08:53.704016 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59fe103c-1de9-48f4-a958-bcd0ce596060-config\") pod \"59fe103c-1de9-48f4-a958-bcd0ce596060\" (UID: \"59fe103c-1de9-48f4-a958-bcd0ce596060\") " Dec 09 10:08:53 crc kubenswrapper[4824]: I1209 10:08:53.704306 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwcnk\" (UniqueName: \"kubernetes.io/projected/59fe103c-1de9-48f4-a958-bcd0ce596060-kube-api-access-lwcnk\") pod \"59fe103c-1de9-48f4-a958-bcd0ce596060\" (UID: \"59fe103c-1de9-48f4-a958-bcd0ce596060\") " Dec 09 10:08:53 crc kubenswrapper[4824]: I1209 10:08:53.704662 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59fe103c-1de9-48f4-a958-bcd0ce596060-config" (OuterVolumeSpecName: "config") pod "59fe103c-1de9-48f4-a958-bcd0ce596060" (UID: "59fe103c-1de9-48f4-a958-bcd0ce596060"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:08:53 crc kubenswrapper[4824]: I1209 10:08:53.705112 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59fe103c-1de9-48f4-a958-bcd0ce596060-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:08:53 crc kubenswrapper[4824]: I1209 10:08:53.708426 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59fe103c-1de9-48f4-a958-bcd0ce596060-kube-api-access-lwcnk" (OuterVolumeSpecName: "kube-api-access-lwcnk") pod "59fe103c-1de9-48f4-a958-bcd0ce596060" (UID: "59fe103c-1de9-48f4-a958-bcd0ce596060"). InnerVolumeSpecName "kube-api-access-lwcnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:08:53 crc kubenswrapper[4824]: I1209 10:08:53.807080 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwcnk\" (UniqueName: \"kubernetes.io/projected/59fe103c-1de9-48f4-a958-bcd0ce596060-kube-api-access-lwcnk\") on node \"crc\" DevicePath \"\"" Dec 09 10:08:54 crc kubenswrapper[4824]: I1209 10:08:54.007668 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerStarted","Data":"3844f695598277b223637109f124ad02a65103550c5696a49ebd000720021e3c"} Dec 09 10:08:54 crc kubenswrapper[4824]: I1209 10:08:54.010990 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ec252d15-4814-478d-ad29-388842d50a32","Type":"ContainerStarted","Data":"ecbb9c2a8fcecabe62ed91b79a214788b05a2f302d5cec36bfafb4429fd5207a"} Dec 09 10:08:54 crc kubenswrapper[4824]: I1209 10:08:54.012708 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c22f05c5-6c0e-4e80-afff-b2642d4a7655","Type":"ContainerStarted","Data":"0dac3e64c1bb4d1fe419d6750fa4b8d78e24e43689628fc4875b72aad46cff92"} Dec 09 10:08:54 crc kubenswrapper[4824]: I1209 10:08:54.020209 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7c85c585b4-tgrff" event={"ID":"e61278e7-d0a8-4039-909d-c2812c8a4a81","Type":"ContainerStarted","Data":"fed6d4a254e7550cda086c8aadaca7770d4dc35060f6273a6c50cd0a7d4c92ae"} Dec 09 10:08:54 crc kubenswrapper[4824]: I1209 10:08:54.021880 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-r2cjc" event={"ID":"59fe103c-1de9-48f4-a958-bcd0ce596060","Type":"ContainerDied","Data":"6e618550834147d803f9373985194cd65d42bf2c723702e1e86bfaace61fe6a3"} Dec 09 10:08:54 crc kubenswrapper[4824]: I1209 10:08:54.022103 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-r2cjc" Dec 09 10:08:54 crc kubenswrapper[4824]: I1209 10:08:54.023441 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rl4nw" event={"ID":"2a54d12b-327d-409e-8652-9525878ae96f","Type":"ContainerStarted","Data":"0d78048d5100ff49d4ee84d6408a228431dcb634fa5fa74e08dcfeb6453fa1ce"} Dec 09 10:08:54 crc kubenswrapper[4824]: I1209 10:08:54.027422 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-lv7r4" event={"ID":"90b0ec71-ccbc-4f8c-8004-2988bcda6655","Type":"ContainerDied","Data":"93fb6c59cc5898d4ff6ce7ccceef8a24f3aa1206fa6be1ca8b152bb091721c8f"} Dec 09 10:08:54 crc kubenswrapper[4824]: I1209 10:08:54.027535 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-lv7r4" Dec 09 10:08:54 crc kubenswrapper[4824]: I1209 10:08:54.030026 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-985bt" event={"ID":"961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d","Type":"ContainerStarted","Data":"03cd7f58b1f6c7fc894a911249c750f8c1a2b4e4b4382cd5802d849d81143d26"} Dec 09 10:08:54 crc kubenswrapper[4824]: I1209 10:08:54.110148 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-r2cjc"] Dec 09 10:08:54 crc kubenswrapper[4824]: I1209 10:08:54.122302 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-r2cjc"] Dec 09 10:08:54 crc kubenswrapper[4824]: I1209 10:08:54.125341 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7c85c585b4-tgrff" podStartSLOduration=31.125317447 podStartE2EDuration="31.125317447s" podCreationTimestamp="2025-12-09 10:08:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:08:54.088603263 +0000 UTC m=+1290.423107930" watchObservedRunningTime="2025-12-09 10:08:54.125317447 +0000 UTC m=+1290.459822114" Dec 09 10:08:54 crc kubenswrapper[4824]: I1209 10:08:54.171002 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-lv7r4"] Dec 09 10:08:54 crc kubenswrapper[4824]: I1209 10:08:54.182484 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-lv7r4"] Dec 09 10:08:55 crc kubenswrapper[4824]: I1209 10:08:55.049899 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"73d6bd70-44c7-4eed-a93a-36df636869cf","Type":"ContainerStarted","Data":"6e0fb2832569ec6c87679f063bf984abcec1863f398d83b73a712b638ae7fc28"} Dec 09 10:08:55 crc kubenswrapper[4824]: I1209 10:08:55.925828 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59fe103c-1de9-48f4-a958-bcd0ce596060" path="/var/lib/kubelet/pods/59fe103c-1de9-48f4-a958-bcd0ce596060/volumes" Dec 09 10:08:55 crc kubenswrapper[4824]: I1209 10:08:55.927195 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90b0ec71-ccbc-4f8c-8004-2988bcda6655" path="/var/lib/kubelet/pods/90b0ec71-ccbc-4f8c-8004-2988bcda6655/volumes" Dec 09 10:08:56 crc kubenswrapper[4824]: I1209 10:08:56.061109 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-8j2s2" event={"ID":"74341961-8a97-42da-9318-4c33280e45e7","Type":"ContainerStarted","Data":"37ee73988422626d6af3078e408f493ca7d12a8c400ef452008110f87c4e8beb"} Dec 09 10:08:56 crc kubenswrapper[4824]: I1209 10:08:56.069684 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"94d75a5f-1730-4993-9948-c42c98719163","Type":"ContainerStarted","Data":"70f793c8069bc8d2b0fd5823635cb4392135e5e442704367b9f8a994600336f5"} Dec 09 10:08:56 crc kubenswrapper[4824]: I1209 10:08:56.069859 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 09 10:08:56 crc kubenswrapper[4824]: I1209 10:08:56.102705 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-8j2s2" podStartSLOduration=16.324122696 podStartE2EDuration="34.102675726s" podCreationTimestamp="2025-12-09 10:08:22 +0000 UTC" firstStartedPulling="2025-12-09 10:08:36.352678725 +0000 UTC m=+1272.687183382" lastFinishedPulling="2025-12-09 10:08:54.131231745 +0000 UTC m=+1290.465736412" observedRunningTime="2025-12-09 10:08:56.076824126 +0000 UTC m=+1292.411328813" watchObservedRunningTime="2025-12-09 10:08:56.102675726 +0000 UTC m=+1292.437180383" Dec 09 10:08:56 crc kubenswrapper[4824]: I1209 10:08:56.129812 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.98718063 podStartE2EDuration="35.129789155s" podCreationTimestamp="2025-12-09 10:08:21 +0000 UTC" firstStartedPulling="2025-12-09 10:08:23.078646602 +0000 UTC m=+1259.413151269" lastFinishedPulling="2025-12-09 10:08:55.221255127 +0000 UTC m=+1291.555759794" observedRunningTime="2025-12-09 10:08:56.101080586 +0000 UTC m=+1292.435585253" watchObservedRunningTime="2025-12-09 10:08:56.129789155 +0000 UTC m=+1292.464293822" Dec 09 10:08:58 crc kubenswrapper[4824]: I1209 10:08:58.090122 4824 generic.go:334] "Generic (PLEG): container finished" podID="73d6bd70-44c7-4eed-a93a-36df636869cf" containerID="6e0fb2832569ec6c87679f063bf984abcec1863f398d83b73a712b638ae7fc28" exitCode=0 Dec 09 10:08:58 crc kubenswrapper[4824]: I1209 10:08:58.090694 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"73d6bd70-44c7-4eed-a93a-36df636869cf","Type":"ContainerDied","Data":"6e0fb2832569ec6c87679f063bf984abcec1863f398d83b73a712b638ae7fc28"} Dec 09 10:08:59 crc kubenswrapper[4824]: I1209 10:08:59.142556 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-985bt" event={"ID":"961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d","Type":"ContainerStarted","Data":"6b72893206670e49372d59e544edb243fc5adca0cb790b7495037f84ec437c8b"} Dec 09 10:08:59 crc kubenswrapper[4824]: I1209 10:08:59.145374 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-985bt" Dec 09 10:08:59 crc kubenswrapper[4824]: I1209 10:08:59.152247 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ec252d15-4814-478d-ad29-388842d50a32","Type":"ContainerStarted","Data":"eee6ee8a5b83850d32773e91999f163baf6553f44eccc7da4278eec06150efd0"} Dec 09 10:08:59 crc kubenswrapper[4824]: I1209 10:08:59.156872 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c22f05c5-6c0e-4e80-afff-b2642d4a7655","Type":"ContainerStarted","Data":"f2fab45580f2249ecec4d19ac43824c635f35f55372c5d67ff9c0ad2be5e7689"} Dec 09 10:08:59 crc kubenswrapper[4824]: I1209 10:08:59.160583 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"73d6bd70-44c7-4eed-a93a-36df636869cf","Type":"ContainerStarted","Data":"3187d927bee90adbfb095796189ae710aa8a824436cd0f34583d4da240bacb9d"} Dec 09 10:08:59 crc kubenswrapper[4824]: I1209 10:08:59.171033 4824 generic.go:334] "Generic (PLEG): container finished" podID="2a54d12b-327d-409e-8652-9525878ae96f" containerID="6bf979d58b1c4d70b5304a17e3044bb09b19f0f269c0f8fe9cef73cc4d4ceea6" exitCode=0 Dec 09 10:08:59 crc kubenswrapper[4824]: I1209 10:08:59.171115 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rl4nw" event={"ID":"2a54d12b-327d-409e-8652-9525878ae96f","Type":"ContainerDied","Data":"6bf979d58b1c4d70b5304a17e3044bb09b19f0f269c0f8fe9cef73cc4d4ceea6"} Dec 09 10:08:59 crc kubenswrapper[4824]: I1209 10:08:59.209573 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-985bt" podStartSLOduration=29.79483702 podStartE2EDuration="34.209534118s" podCreationTimestamp="2025-12-09 10:08:25 +0000 UTC" firstStartedPulling="2025-12-09 10:08:53.488116249 +0000 UTC m=+1289.822620916" lastFinishedPulling="2025-12-09 10:08:57.902813347 +0000 UTC m=+1294.237318014" observedRunningTime="2025-12-09 10:08:59.170994377 +0000 UTC m=+1295.505499044" watchObservedRunningTime="2025-12-09 10:08:59.209534118 +0000 UTC m=+1295.544038785" Dec 09 10:08:59 crc kubenswrapper[4824]: I1209 10:08:59.210033 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=11.083868549 podStartE2EDuration="42.210026524s" podCreationTimestamp="2025-12-09 10:08:17 +0000 UTC" firstStartedPulling="2025-12-09 10:08:21.375705311 +0000 UTC m=+1257.710209978" lastFinishedPulling="2025-12-09 10:08:52.501863286 +0000 UTC m=+1288.836367953" observedRunningTime="2025-12-09 10:08:59.19412161 +0000 UTC m=+1295.528626277" watchObservedRunningTime="2025-12-09 10:08:59.210026524 +0000 UTC m=+1295.544531191" Dec 09 10:08:59 crc kubenswrapper[4824]: I1209 10:08:59.348713 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 09 10:08:59 crc kubenswrapper[4824]: I1209 10:08:59.349190 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 09 10:09:00 crc kubenswrapper[4824]: I1209 10:09:00.183438 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"07c1822c-1257-4d8c-9a17-d921929000fa","Type":"ContainerStarted","Data":"b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c"} Dec 09 10:09:00 crc kubenswrapper[4824]: I1209 10:09:00.189210 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rl4nw" event={"ID":"2a54d12b-327d-409e-8652-9525878ae96f","Type":"ContainerStarted","Data":"59ec82b41a8674a766eefe22dbaac5e0067fc1515eef7c51d9008a28a00bb8f7"} Dec 09 10:09:00 crc kubenswrapper[4824]: I1209 10:09:00.189281 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rl4nw" event={"ID":"2a54d12b-327d-409e-8652-9525878ae96f","Type":"ContainerStarted","Data":"30b330b9e5d2c3fec1fefb020d17fdb3949c8b53fccf44d1433b8104f2e7ed51"} Dec 09 10:09:00 crc kubenswrapper[4824]: I1209 10:09:00.189355 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-rl4nw" Dec 09 10:09:00 crc kubenswrapper[4824]: I1209 10:09:00.192775 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"56148e0f-636f-410e-bfb5-342da01e8c76","Type":"ContainerStarted","Data":"1c78cc837b39b1d11f13234c33e6e9cc8128107fd0e4df95650a8b44ff42d53a"} Dec 09 10:09:00 crc kubenswrapper[4824]: I1209 10:09:00.262618 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-rl4nw" podStartSLOduration=30.860283522 podStartE2EDuration="35.262599639s" podCreationTimestamp="2025-12-09 10:08:25 +0000 UTC" firstStartedPulling="2025-12-09 10:08:53.494956115 +0000 UTC m=+1289.829460782" lastFinishedPulling="2025-12-09 10:08:57.897272222 +0000 UTC m=+1294.231776899" observedRunningTime="2025-12-09 10:09:00.261125652 +0000 UTC m=+1296.595630339" watchObservedRunningTime="2025-12-09 10:09:00.262599639 +0000 UTC m=+1296.597104306" Dec 09 10:09:00 crc kubenswrapper[4824]: I1209 10:09:00.444184 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-rl4nw" Dec 09 10:09:01 crc kubenswrapper[4824]: I1209 10:09:01.692294 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 09 10:09:02 crc kubenswrapper[4824]: I1209 10:09:02.215141 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ec252d15-4814-478d-ad29-388842d50a32","Type":"ContainerStarted","Data":"7fa8e80e99d597d212a823df2824b5c493d7e49b2158750d7344f5b8413dff27"} Dec 09 10:09:02 crc kubenswrapper[4824]: I1209 10:09:02.220723 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c22f05c5-6c0e-4e80-afff-b2642d4a7655","Type":"ContainerStarted","Data":"d619234f2fc3c345f4f5dfe4e2ca07d4385ecd259de2c969fc154c99da6b4f0b"} Dec 09 10:09:02 crc kubenswrapper[4824]: I1209 10:09:02.238982 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=27.360778782 podStartE2EDuration="35.238966886s" podCreationTimestamp="2025-12-09 10:08:27 +0000 UTC" firstStartedPulling="2025-12-09 10:08:53.486327403 +0000 UTC m=+1289.820832070" lastFinishedPulling="2025-12-09 10:09:01.364515507 +0000 UTC m=+1297.699020174" observedRunningTime="2025-12-09 10:09:02.236405205 +0000 UTC m=+1298.570909882" watchObservedRunningTime="2025-12-09 10:09:02.238966886 +0000 UTC m=+1298.573471543" Dec 09 10:09:02 crc kubenswrapper[4824]: I1209 10:09:02.260601 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=32.461841564 podStartE2EDuration="40.260578111s" podCreationTimestamp="2025-12-09 10:08:22 +0000 UTC" firstStartedPulling="2025-12-09 10:08:53.557908251 +0000 UTC m=+1289.892412918" lastFinishedPulling="2025-12-09 10:09:01.356644798 +0000 UTC m=+1297.691149465" observedRunningTime="2025-12-09 10:09:02.256326837 +0000 UTC m=+1298.590831504" watchObservedRunningTime="2025-12-09 10:09:02.260578111 +0000 UTC m=+1298.595082778" Dec 09 10:09:02 crc kubenswrapper[4824]: I1209 10:09:02.352294 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 09 10:09:02 crc kubenswrapper[4824]: I1209 10:09:02.395340 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.234462 4824 generic.go:334] "Generic (PLEG): container finished" podID="56148e0f-636f-410e-bfb5-342da01e8c76" containerID="1c78cc837b39b1d11f13234c33e6e9cc8128107fd0e4df95650a8b44ff42d53a" exitCode=0 Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.235108 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"56148e0f-636f-410e-bfb5-342da01e8c76","Type":"ContainerDied","Data":"1c78cc837b39b1d11f13234c33e6e9cc8128107fd0e4df95650a8b44ff42d53a"} Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.235361 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.288107 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.424021 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.438034 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.540462 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.581488 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-fhkvj"] Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.625363 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-p4czz"] Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.628511 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.634653 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.642225 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-p4czz"] Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.703285 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.716964 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.721502 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.745557 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-6f874"] Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.748677 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-6f874" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.756372 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.773129 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-6f874"] Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.785587 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.786193 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9177fe98-0bd2-43b4-85bb-f9218d48211e-config\") pod \"dnsmasq-dns-7f896c8c65-p4czz\" (UID: \"9177fe98-0bd2-43b4-85bb-f9218d48211e\") " pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.786385 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78pwt\" (UniqueName: \"kubernetes.io/projected/9177fe98-0bd2-43b4-85bb-f9218d48211e-kube-api-access-78pwt\") pod \"dnsmasq-dns-7f896c8c65-p4czz\" (UID: \"9177fe98-0bd2-43b4-85bb-f9218d48211e\") " pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.786547 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9177fe98-0bd2-43b4-85bb-f9218d48211e-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-p4czz\" (UID: \"9177fe98-0bd2-43b4-85bb-f9218d48211e\") " pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.786658 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9177fe98-0bd2-43b4-85bb-f9218d48211e-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-p4czz\" (UID: \"9177fe98-0bd2-43b4-85bb-f9218d48211e\") " pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.897512 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/c0ec6983-7e9d-449f-99ce-44b0bc708f19-ovn-rundir\") pod \"ovn-controller-metrics-6f874\" (UID: \"c0ec6983-7e9d-449f-99ce-44b0bc708f19\") " pod="openstack/ovn-controller-metrics-6f874" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.898114 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9177fe98-0bd2-43b4-85bb-f9218d48211e-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-p4czz\" (UID: \"9177fe98-0bd2-43b4-85bb-f9218d48211e\") " pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.898179 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0ec6983-7e9d-449f-99ce-44b0bc708f19-config\") pod \"ovn-controller-metrics-6f874\" (UID: \"c0ec6983-7e9d-449f-99ce-44b0bc708f19\") " pod="openstack/ovn-controller-metrics-6f874" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.898252 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9177fe98-0bd2-43b4-85bb-f9218d48211e-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-p4czz\" (UID: \"9177fe98-0bd2-43b4-85bb-f9218d48211e\") " pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.898406 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9177fe98-0bd2-43b4-85bb-f9218d48211e-config\") pod \"dnsmasq-dns-7f896c8c65-p4czz\" (UID: \"9177fe98-0bd2-43b4-85bb-f9218d48211e\") " pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.898452 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/c0ec6983-7e9d-449f-99ce-44b0bc708f19-ovs-rundir\") pod \"ovn-controller-metrics-6f874\" (UID: \"c0ec6983-7e9d-449f-99ce-44b0bc708f19\") " pod="openstack/ovn-controller-metrics-6f874" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.898480 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0ec6983-7e9d-449f-99ce-44b0bc708f19-combined-ca-bundle\") pod \"ovn-controller-metrics-6f874\" (UID: \"c0ec6983-7e9d-449f-99ce-44b0bc708f19\") " pod="openstack/ovn-controller-metrics-6f874" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.898540 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0ec6983-7e9d-449f-99ce-44b0bc708f19-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-6f874\" (UID: \"c0ec6983-7e9d-449f-99ce-44b0bc708f19\") " pod="openstack/ovn-controller-metrics-6f874" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.902108 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9177fe98-0bd2-43b4-85bb-f9218d48211e-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-p4czz\" (UID: \"9177fe98-0bd2-43b4-85bb-f9218d48211e\") " pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.902304 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9177fe98-0bd2-43b4-85bb-f9218d48211e-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-p4czz\" (UID: \"9177fe98-0bd2-43b4-85bb-f9218d48211e\") " pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.902330 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9177fe98-0bd2-43b4-85bb-f9218d48211e-config\") pod \"dnsmasq-dns-7f896c8c65-p4czz\" (UID: \"9177fe98-0bd2-43b4-85bb-f9218d48211e\") " pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.910820 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g774m\" (UniqueName: \"kubernetes.io/projected/c0ec6983-7e9d-449f-99ce-44b0bc708f19-kube-api-access-g774m\") pod \"ovn-controller-metrics-6f874\" (UID: \"c0ec6983-7e9d-449f-99ce-44b0bc708f19\") " pod="openstack/ovn-controller-metrics-6f874" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.910949 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78pwt\" (UniqueName: \"kubernetes.io/projected/9177fe98-0bd2-43b4-85bb-f9218d48211e-kube-api-access-78pwt\") pod \"dnsmasq-dns-7f896c8c65-p4czz\" (UID: \"9177fe98-0bd2-43b4-85bb-f9218d48211e\") " pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" Dec 09 10:09:03 crc kubenswrapper[4824]: I1209 10:09:03.967260 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78pwt\" (UniqueName: \"kubernetes.io/projected/9177fe98-0bd2-43b4-85bb-f9218d48211e-kube-api-access-78pwt\") pod \"dnsmasq-dns-7f896c8c65-p4czz\" (UID: \"9177fe98-0bd2-43b4-85bb-f9218d48211e\") " pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.021372 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0ec6983-7e9d-449f-99ce-44b0bc708f19-config\") pod \"ovn-controller-metrics-6f874\" (UID: \"c0ec6983-7e9d-449f-99ce-44b0bc708f19\") " pod="openstack/ovn-controller-metrics-6f874" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.021727 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/c0ec6983-7e9d-449f-99ce-44b0bc708f19-ovs-rundir\") pod \"ovn-controller-metrics-6f874\" (UID: \"c0ec6983-7e9d-449f-99ce-44b0bc708f19\") " pod="openstack/ovn-controller-metrics-6f874" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.021760 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0ec6983-7e9d-449f-99ce-44b0bc708f19-combined-ca-bundle\") pod \"ovn-controller-metrics-6f874\" (UID: \"c0ec6983-7e9d-449f-99ce-44b0bc708f19\") " pod="openstack/ovn-controller-metrics-6f874" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.021854 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0ec6983-7e9d-449f-99ce-44b0bc708f19-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-6f874\" (UID: \"c0ec6983-7e9d-449f-99ce-44b0bc708f19\") " pod="openstack/ovn-controller-metrics-6f874" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.021933 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g774m\" (UniqueName: \"kubernetes.io/projected/c0ec6983-7e9d-449f-99ce-44b0bc708f19-kube-api-access-g774m\") pod \"ovn-controller-metrics-6f874\" (UID: \"c0ec6983-7e9d-449f-99ce-44b0bc708f19\") " pod="openstack/ovn-controller-metrics-6f874" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.022052 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/c0ec6983-7e9d-449f-99ce-44b0bc708f19-ovn-rundir\") pod \"ovn-controller-metrics-6f874\" (UID: \"c0ec6983-7e9d-449f-99ce-44b0bc708f19\") " pod="openstack/ovn-controller-metrics-6f874" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.022580 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/c0ec6983-7e9d-449f-99ce-44b0bc708f19-ovn-rundir\") pod \"ovn-controller-metrics-6f874\" (UID: \"c0ec6983-7e9d-449f-99ce-44b0bc708f19\") " pod="openstack/ovn-controller-metrics-6f874" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.022649 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/c0ec6983-7e9d-449f-99ce-44b0bc708f19-ovs-rundir\") pod \"ovn-controller-metrics-6f874\" (UID: \"c0ec6983-7e9d-449f-99ce-44b0bc708f19\") " pod="openstack/ovn-controller-metrics-6f874" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.023374 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0ec6983-7e9d-449f-99ce-44b0bc708f19-config\") pod \"ovn-controller-metrics-6f874\" (UID: \"c0ec6983-7e9d-449f-99ce-44b0bc708f19\") " pod="openstack/ovn-controller-metrics-6f874" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.040407 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0ec6983-7e9d-449f-99ce-44b0bc708f19-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-6f874\" (UID: \"c0ec6983-7e9d-449f-99ce-44b0bc708f19\") " pod="openstack/ovn-controller-metrics-6f874" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.041307 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0ec6983-7e9d-449f-99ce-44b0bc708f19-combined-ca-bundle\") pod \"ovn-controller-metrics-6f874\" (UID: \"c0ec6983-7e9d-449f-99ce-44b0bc708f19\") " pod="openstack/ovn-controller-metrics-6f874" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.051445 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-cgdx9"] Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.060041 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g774m\" (UniqueName: \"kubernetes.io/projected/c0ec6983-7e9d-449f-99ce-44b0bc708f19-kube-api-access-g774m\") pod \"ovn-controller-metrics-6f874\" (UID: \"c0ec6983-7e9d-449f-99ce-44b0bc708f19\") " pod="openstack/ovn-controller-metrics-6f874" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.092010 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-6f874" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.099432 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-fhkvj" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.103846 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-z775c"] Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.107382 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-z775c" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.109819 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.129113 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-z775c"] Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.230643 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n2kp\" (UniqueName: \"kubernetes.io/projected/a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae-kube-api-access-9n2kp\") pod \"a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae\" (UID: \"a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae\") " Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.233902 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae-config\") pod \"a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae\" (UID: \"a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae\") " Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.234042 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae-dns-svc\") pod \"a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae\" (UID: \"a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae\") " Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.234804 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae" (UID: "a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.234935 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae-config" (OuterVolumeSpecName: "config") pod "a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae" (UID: "a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.236657 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n7h2\" (UniqueName: \"kubernetes.io/projected/ddc24cb1-2724-40b8-bfc9-81474d67d928-kube-api-access-2n7h2\") pod \"dnsmasq-dns-86db49b7ff-z775c\" (UID: \"ddc24cb1-2724-40b8-bfc9-81474d67d928\") " pod="openstack/dnsmasq-dns-86db49b7ff-z775c" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.236970 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ddc24cb1-2724-40b8-bfc9-81474d67d928-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-z775c\" (UID: \"ddc24cb1-2724-40b8-bfc9-81474d67d928\") " pod="openstack/dnsmasq-dns-86db49b7ff-z775c" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.237022 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddc24cb1-2724-40b8-bfc9-81474d67d928-config\") pod \"dnsmasq-dns-86db49b7ff-z775c\" (UID: \"ddc24cb1-2724-40b8-bfc9-81474d67d928\") " pod="openstack/dnsmasq-dns-86db49b7ff-z775c" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.237068 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ddc24cb1-2724-40b8-bfc9-81474d67d928-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-z775c\" (UID: \"ddc24cb1-2724-40b8-bfc9-81474d67d928\") " pod="openstack/dnsmasq-dns-86db49b7ff-z775c" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.237182 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ddc24cb1-2724-40b8-bfc9-81474d67d928-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-z775c\" (UID: \"ddc24cb1-2724-40b8-bfc9-81474d67d928\") " pod="openstack/dnsmasq-dns-86db49b7ff-z775c" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.237400 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.237416 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.250961 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae-kube-api-access-9n2kp" (OuterVolumeSpecName: "kube-api-access-9n2kp") pod "a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae" (UID: "a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae"). InnerVolumeSpecName "kube-api-access-9n2kp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.265317 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.285815 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-fhkvj" event={"ID":"a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae","Type":"ContainerDied","Data":"2baadce7f2404f87cc1cc9f11b0fc8dbc2f4b841f052433b72ece8c232c9e4b7"} Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.285944 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-fhkvj" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.298030 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"56148e0f-636f-410e-bfb5-342da01e8c76","Type":"ContainerStarted","Data":"4f85804d9b180ab6d108fd3769e0a57f89c384076fef52095c519bf02c5c868e"} Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.300474 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.305540 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.339645 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ddc24cb1-2724-40b8-bfc9-81474d67d928-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-z775c\" (UID: \"ddc24cb1-2724-40b8-bfc9-81474d67d928\") " pod="openstack/dnsmasq-dns-86db49b7ff-z775c" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.339710 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddc24cb1-2724-40b8-bfc9-81474d67d928-config\") pod \"dnsmasq-dns-86db49b7ff-z775c\" (UID: \"ddc24cb1-2724-40b8-bfc9-81474d67d928\") " pod="openstack/dnsmasq-dns-86db49b7ff-z775c" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.339746 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ddc24cb1-2724-40b8-bfc9-81474d67d928-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-z775c\" (UID: \"ddc24cb1-2724-40b8-bfc9-81474d67d928\") " pod="openstack/dnsmasq-dns-86db49b7ff-z775c" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.339829 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ddc24cb1-2724-40b8-bfc9-81474d67d928-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-z775c\" (UID: \"ddc24cb1-2724-40b8-bfc9-81474d67d928\") " pod="openstack/dnsmasq-dns-86db49b7ff-z775c" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.340137 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n7h2\" (UniqueName: \"kubernetes.io/projected/ddc24cb1-2724-40b8-bfc9-81474d67d928-kube-api-access-2n7h2\") pod \"dnsmasq-dns-86db49b7ff-z775c\" (UID: \"ddc24cb1-2724-40b8-bfc9-81474d67d928\") " pod="openstack/dnsmasq-dns-86db49b7ff-z775c" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.340310 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n2kp\" (UniqueName: \"kubernetes.io/projected/a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae-kube-api-access-9n2kp\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.342151 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ddc24cb1-2724-40b8-bfc9-81474d67d928-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-z775c\" (UID: \"ddc24cb1-2724-40b8-bfc9-81474d67d928\") " pod="openstack/dnsmasq-dns-86db49b7ff-z775c" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.343308 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ddc24cb1-2724-40b8-bfc9-81474d67d928-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-z775c\" (UID: \"ddc24cb1-2724-40b8-bfc9-81474d67d928\") " pod="openstack/dnsmasq-dns-86db49b7ff-z775c" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.343834 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ddc24cb1-2724-40b8-bfc9-81474d67d928-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-z775c\" (UID: \"ddc24cb1-2724-40b8-bfc9-81474d67d928\") " pod="openstack/dnsmasq-dns-86db49b7ff-z775c" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.346852 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddc24cb1-2724-40b8-bfc9-81474d67d928-config\") pod \"dnsmasq-dns-86db49b7ff-z775c\" (UID: \"ddc24cb1-2724-40b8-bfc9-81474d67d928\") " pod="openstack/dnsmasq-dns-86db49b7ff-z775c" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.357613 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371987.49719 podStartE2EDuration="49.357584642s" podCreationTimestamp="2025-12-09 10:08:15 +0000 UTC" firstStartedPulling="2025-12-09 10:08:18.286877011 +0000 UTC m=+1254.621381678" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:09:04.330539415 +0000 UTC m=+1300.665044102" watchObservedRunningTime="2025-12-09 10:09:04.357584642 +0000 UTC m=+1300.692089329" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.385275 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n7h2\" (UniqueName: \"kubernetes.io/projected/ddc24cb1-2724-40b8-bfc9-81474d67d928-kube-api-access-2n7h2\") pod \"dnsmasq-dns-86db49b7ff-z775c\" (UID: \"ddc24cb1-2724-40b8-bfc9-81474d67d928\") " pod="openstack/dnsmasq-dns-86db49b7ff-z775c" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.413355 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-77dcf56988-s8mzf"] Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.414493 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.457770 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-z775c" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.566087 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-fhkvj"] Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.578908 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-fhkvj"] Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.731059 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-cgdx9" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.752800 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n79b5\" (UniqueName: \"kubernetes.io/projected/c51bee9b-4c7f-4809-8406-488db1c6223a-kube-api-access-n79b5\") pod \"c51bee9b-4c7f-4809-8406-488db1c6223a\" (UID: \"c51bee9b-4c7f-4809-8406-488db1c6223a\") " Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.753135 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c51bee9b-4c7f-4809-8406-488db1c6223a-config\") pod \"c51bee9b-4c7f-4809-8406-488db1c6223a\" (UID: \"c51bee9b-4c7f-4809-8406-488db1c6223a\") " Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.753208 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c51bee9b-4c7f-4809-8406-488db1c6223a-dns-svc\") pod \"c51bee9b-4c7f-4809-8406-488db1c6223a\" (UID: \"c51bee9b-4c7f-4809-8406-488db1c6223a\") " Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.754368 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c51bee9b-4c7f-4809-8406-488db1c6223a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c51bee9b-4c7f-4809-8406-488db1c6223a" (UID: "c51bee9b-4c7f-4809-8406-488db1c6223a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.765491 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c51bee9b-4c7f-4809-8406-488db1c6223a-config" (OuterVolumeSpecName: "config") pod "c51bee9b-4c7f-4809-8406-488db1c6223a" (UID: "c51bee9b-4c7f-4809-8406-488db1c6223a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.774086 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c51bee9b-4c7f-4809-8406-488db1c6223a-kube-api-access-n79b5" (OuterVolumeSpecName: "kube-api-access-n79b5") pod "c51bee9b-4c7f-4809-8406-488db1c6223a" (UID: "c51bee9b-4c7f-4809-8406-488db1c6223a"). InnerVolumeSpecName "kube-api-access-n79b5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.837304 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.839898 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.844410 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.849507 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.849821 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.850100 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-7mdw6" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.872669 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2dd0680-d32f-45e8-b834-6a327d2eaa21-config\") pod \"ovn-northd-0\" (UID: \"c2dd0680-d32f-45e8-b834-6a327d2eaa21\") " pod="openstack/ovn-northd-0" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.872880 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c2dd0680-d32f-45e8-b834-6a327d2eaa21-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c2dd0680-d32f-45e8-b834-6a327d2eaa21\") " pod="openstack/ovn-northd-0" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.872936 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2dd0680-d32f-45e8-b834-6a327d2eaa21-scripts\") pod \"ovn-northd-0\" (UID: \"c2dd0680-d32f-45e8-b834-6a327d2eaa21\") " pod="openstack/ovn-northd-0" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.872973 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2dd0680-d32f-45e8-b834-6a327d2eaa21-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c2dd0680-d32f-45e8-b834-6a327d2eaa21\") " pod="openstack/ovn-northd-0" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.873044 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2dd0680-d32f-45e8-b834-6a327d2eaa21-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c2dd0680-d32f-45e8-b834-6a327d2eaa21\") " pod="openstack/ovn-northd-0" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.873176 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8czds\" (UniqueName: \"kubernetes.io/projected/c2dd0680-d32f-45e8-b834-6a327d2eaa21-kube-api-access-8czds\") pod \"ovn-northd-0\" (UID: \"c2dd0680-d32f-45e8-b834-6a327d2eaa21\") " pod="openstack/ovn-northd-0" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.873227 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2dd0680-d32f-45e8-b834-6a327d2eaa21-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c2dd0680-d32f-45e8-b834-6a327d2eaa21\") " pod="openstack/ovn-northd-0" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.873361 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c51bee9b-4c7f-4809-8406-488db1c6223a-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.873376 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c51bee9b-4c7f-4809-8406-488db1c6223a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:04 crc kubenswrapper[4824]: I1209 10:09:04.873389 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n79b5\" (UniqueName: \"kubernetes.io/projected/c51bee9b-4c7f-4809-8406-488db1c6223a-kube-api-access-n79b5\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:05 crc kubenswrapper[4824]: I1209 10:09:05.001455 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-6f874"] Dec 09 10:09:05 crc kubenswrapper[4824]: I1209 10:09:05.011576 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2dd0680-d32f-45e8-b834-6a327d2eaa21-config\") pod \"ovn-northd-0\" (UID: \"c2dd0680-d32f-45e8-b834-6a327d2eaa21\") " pod="openstack/ovn-northd-0" Dec 09 10:09:05 crc kubenswrapper[4824]: I1209 10:09:05.011901 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c2dd0680-d32f-45e8-b834-6a327d2eaa21-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c2dd0680-d32f-45e8-b834-6a327d2eaa21\") " pod="openstack/ovn-northd-0" Dec 09 10:09:05 crc kubenswrapper[4824]: I1209 10:09:05.014154 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2dd0680-d32f-45e8-b834-6a327d2eaa21-config\") pod \"ovn-northd-0\" (UID: \"c2dd0680-d32f-45e8-b834-6a327d2eaa21\") " pod="openstack/ovn-northd-0" Dec 09 10:09:05 crc kubenswrapper[4824]: I1209 10:09:05.011980 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2dd0680-d32f-45e8-b834-6a327d2eaa21-scripts\") pod \"ovn-northd-0\" (UID: \"c2dd0680-d32f-45e8-b834-6a327d2eaa21\") " pod="openstack/ovn-northd-0" Dec 09 10:09:05 crc kubenswrapper[4824]: I1209 10:09:05.016513 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2dd0680-d32f-45e8-b834-6a327d2eaa21-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c2dd0680-d32f-45e8-b834-6a327d2eaa21\") " pod="openstack/ovn-northd-0" Dec 09 10:09:05 crc kubenswrapper[4824]: I1209 10:09:05.016675 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2dd0680-d32f-45e8-b834-6a327d2eaa21-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c2dd0680-d32f-45e8-b834-6a327d2eaa21\") " pod="openstack/ovn-northd-0" Dec 09 10:09:05 crc kubenswrapper[4824]: I1209 10:09:05.016856 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8czds\" (UniqueName: \"kubernetes.io/projected/c2dd0680-d32f-45e8-b834-6a327d2eaa21-kube-api-access-8czds\") pod \"ovn-northd-0\" (UID: \"c2dd0680-d32f-45e8-b834-6a327d2eaa21\") " pod="openstack/ovn-northd-0" Dec 09 10:09:05 crc kubenswrapper[4824]: I1209 10:09:05.016944 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2dd0680-d32f-45e8-b834-6a327d2eaa21-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c2dd0680-d32f-45e8-b834-6a327d2eaa21\") " pod="openstack/ovn-northd-0" Dec 09 10:09:05 crc kubenswrapper[4824]: I1209 10:09:05.017487 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2dd0680-d32f-45e8-b834-6a327d2eaa21-scripts\") pod \"ovn-northd-0\" (UID: \"c2dd0680-d32f-45e8-b834-6a327d2eaa21\") " pod="openstack/ovn-northd-0" Dec 09 10:09:05 crc kubenswrapper[4824]: I1209 10:09:05.023352 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2dd0680-d32f-45e8-b834-6a327d2eaa21-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c2dd0680-d32f-45e8-b834-6a327d2eaa21\") " pod="openstack/ovn-northd-0" Dec 09 10:09:05 crc kubenswrapper[4824]: I1209 10:09:05.023970 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c2dd0680-d32f-45e8-b834-6a327d2eaa21-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c2dd0680-d32f-45e8-b834-6a327d2eaa21\") " pod="openstack/ovn-northd-0" Dec 09 10:09:05 crc kubenswrapper[4824]: I1209 10:09:05.028827 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2dd0680-d32f-45e8-b834-6a327d2eaa21-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c2dd0680-d32f-45e8-b834-6a327d2eaa21\") " pod="openstack/ovn-northd-0" Dec 09 10:09:05 crc kubenswrapper[4824]: I1209 10:09:05.033039 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2dd0680-d32f-45e8-b834-6a327d2eaa21-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c2dd0680-d32f-45e8-b834-6a327d2eaa21\") " pod="openstack/ovn-northd-0" Dec 09 10:09:05 crc kubenswrapper[4824]: I1209 10:09:05.070110 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8czds\" (UniqueName: \"kubernetes.io/projected/c2dd0680-d32f-45e8-b834-6a327d2eaa21-kube-api-access-8czds\") pod \"ovn-northd-0\" (UID: \"c2dd0680-d32f-45e8-b834-6a327d2eaa21\") " pod="openstack/ovn-northd-0" Dec 09 10:09:05 crc kubenswrapper[4824]: I1209 10:09:05.100758 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 09 10:09:05 crc kubenswrapper[4824]: I1209 10:09:05.103973 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 09 10:09:05 crc kubenswrapper[4824]: I1209 10:09:05.364302 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-cgdx9" event={"ID":"c51bee9b-4c7f-4809-8406-488db1c6223a","Type":"ContainerDied","Data":"a167c5aec82479abbf9724d7afb2e62edd060fdf30226a87aa6fdcd62a3a3161"} Dec 09 10:09:05 crc kubenswrapper[4824]: I1209 10:09:05.364443 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-cgdx9" Dec 09 10:09:05 crc kubenswrapper[4824]: I1209 10:09:05.381603 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-6f874" event={"ID":"c0ec6983-7e9d-449f-99ce-44b0bc708f19","Type":"ContainerStarted","Data":"fabca8be955f45ab596b3ded0597ae8d0dde9956b7d144534f4a23cee7ba3850"} Dec 09 10:09:05 crc kubenswrapper[4824]: I1209 10:09:05.923271 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae" path="/var/lib/kubelet/pods/a6311e5c-20b6-47c5-8d8a-f540cfe8a3ae/volumes" Dec 09 10:09:06 crc kubenswrapper[4824]: I1209 10:09:06.507348 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-p4czz"] Dec 09 10:09:06 crc kubenswrapper[4824]: I1209 10:09:06.520489 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-z775c"] Dec 09 10:09:06 crc kubenswrapper[4824]: I1209 10:09:06.584661 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-cgdx9"] Dec 09 10:09:06 crc kubenswrapper[4824]: I1209 10:09:06.595332 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-cgdx9"] Dec 09 10:09:06 crc kubenswrapper[4824]: I1209 10:09:06.959697 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 09 10:09:06 crc kubenswrapper[4824]: W1209 10:09:06.965332 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2dd0680_d32f_45e8_b834_6a327d2eaa21.slice/crio-0e12c5ba0e37438210ea9fc36ce123372e321265d487ca10ec0bcc0deea4ec9f WatchSource:0}: Error finding container 0e12c5ba0e37438210ea9fc36ce123372e321265d487ca10ec0bcc0deea4ec9f: Status 404 returned error can't find the container with id 0e12c5ba0e37438210ea9fc36ce123372e321265d487ca10ec0bcc0deea4ec9f Dec 09 10:09:07 crc kubenswrapper[4824]: I1209 10:09:07.527095 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" event={"ID":"9177fe98-0bd2-43b4-85bb-f9218d48211e","Type":"ContainerStarted","Data":"4b087d5395fa6b4a6d06e08ddb1513837e0502acc753cd2a79cd585626b29941"} Dec 09 10:09:07 crc kubenswrapper[4824]: I1209 10:09:07.529418 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c2dd0680-d32f-45e8-b834-6a327d2eaa21","Type":"ContainerStarted","Data":"0e12c5ba0e37438210ea9fc36ce123372e321265d487ca10ec0bcc0deea4ec9f"} Dec 09 10:09:07 crc kubenswrapper[4824]: I1209 10:09:07.533531 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-6f874" event={"ID":"c0ec6983-7e9d-449f-99ce-44b0bc708f19","Type":"ContainerStarted","Data":"bcbd9a5c9add0a7ea8c14311928de644a69322b9205b2c26848e696f7a7c498d"} Dec 09 10:09:07 crc kubenswrapper[4824]: I1209 10:09:07.536507 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-z775c" event={"ID":"ddc24cb1-2724-40b8-bfc9-81474d67d928","Type":"ContainerStarted","Data":"8c48407177b665d20502ea8c60de2bbdab77c7310c46231bab2151641bfb905a"} Dec 09 10:09:07 crc kubenswrapper[4824]: I1209 10:09:07.566149 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-6f874" podStartSLOduration=4.5661226070000005 podStartE2EDuration="4.566122607s" podCreationTimestamp="2025-12-09 10:09:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:09:07.552451024 +0000 UTC m=+1303.886955701" watchObservedRunningTime="2025-12-09 10:09:07.566122607 +0000 UTC m=+1303.900627274" Dec 09 10:09:07 crc kubenswrapper[4824]: I1209 10:09:07.744706 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 09 10:09:07 crc kubenswrapper[4824]: I1209 10:09:07.744774 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 09 10:09:07 crc kubenswrapper[4824]: I1209 10:09:07.923061 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c51bee9b-4c7f-4809-8406-488db1c6223a" path="/var/lib/kubelet/pods/c51bee9b-4c7f-4809-8406-488db1c6223a/volumes" Dec 09 10:09:10 crc kubenswrapper[4824]: I1209 10:09:10.016989 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"4afe9ced-d6b6-4684-96f3-a9a419711cc7","Type":"ContainerStarted","Data":"a653b6e32d1363e3973a64d0df688f0fc3ecac54cfa459b9889cd8146982a045"} Dec 09 10:09:10 crc kubenswrapper[4824]: I1209 10:09:10.019142 4824 generic.go:334] "Generic (PLEG): container finished" podID="ddc24cb1-2724-40b8-bfc9-81474d67d928" containerID="31ebf1b50662f71b55fc3e7f1aa08e494f7f99302d8044c600a3e27be0d119a6" exitCode=0 Dec 09 10:09:10 crc kubenswrapper[4824]: I1209 10:09:10.019191 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-z775c" event={"ID":"ddc24cb1-2724-40b8-bfc9-81474d67d928","Type":"ContainerDied","Data":"31ebf1b50662f71b55fc3e7f1aa08e494f7f99302d8044c600a3e27be0d119a6"} Dec 09 10:09:10 crc kubenswrapper[4824]: I1209 10:09:10.024479 4824 generic.go:334] "Generic (PLEG): container finished" podID="9177fe98-0bd2-43b4-85bb-f9218d48211e" containerID="3516d3c603a4d43ca10f7c358d86e94b214fe1665395df66d0bbae876f8b963b" exitCode=0 Dec 09 10:09:10 crc kubenswrapper[4824]: I1209 10:09:10.024588 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" event={"ID":"9177fe98-0bd2-43b4-85bb-f9218d48211e","Type":"ContainerDied","Data":"3516d3c603a4d43ca10f7c358d86e94b214fe1665395df66d0bbae876f8b963b"} Dec 09 10:09:10 crc kubenswrapper[4824]: I1209 10:09:10.029060 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55","Type":"ContainerStarted","Data":"17519c5dfbefe63576e185d8b281c4f8a3c5c8eaff5cd9a05b5db8801ca6f9b2"} Dec 09 10:09:11 crc kubenswrapper[4824]: I1209 10:09:11.838042 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-index-2jnrg" podUID="25c6995d-6a34-43fc-9b1b-6c34ed9aec46" containerName="registry-server" probeResult="failure" output=< Dec 09 10:09:11 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 10:09:11 crc kubenswrapper[4824]: > Dec 09 10:09:13 crc kubenswrapper[4824]: I1209 10:09:13.948270 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 09 10:09:13 crc kubenswrapper[4824]: I1209 10:09:13.971342 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=10.337835669 podStartE2EDuration="55.97131784s" podCreationTimestamp="2025-12-09 10:08:18 +0000 UTC" firstStartedPulling="2025-12-09 10:08:21.94887648 +0000 UTC m=+1258.283381137" lastFinishedPulling="2025-12-09 10:09:07.582358651 +0000 UTC m=+1303.916863308" observedRunningTime="2025-12-09 10:09:13.963857874 +0000 UTC m=+1310.298362541" watchObservedRunningTime="2025-12-09 10:09:13.97131784 +0000 UTC m=+1310.305822507" Dec 09 10:09:14 crc kubenswrapper[4824]: I1209 10:09:14.961134 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 09 10:09:16 crc kubenswrapper[4824]: I1209 10:09:16.977679 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c2dd0680-d32f-45e8-b834-6a327d2eaa21","Type":"ContainerStarted","Data":"a13a50a9012203c512c34fec894886dfe2d054768904caa6f9ab13fa1fc56c87"} Dec 09 10:09:17 crc kubenswrapper[4824]: I1209 10:09:17.990253 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8edb90f3-f84b-4700-993f-21ffa950b620","Type":"ContainerStarted","Data":"7cf4f49e8f8e0800f59be8f22816ba55ee3205fc0e8f6c81552552f0e44a23f7"} Dec 09 10:09:17 crc kubenswrapper[4824]: I1209 10:09:17.993759 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-z775c" event={"ID":"ddc24cb1-2724-40b8-bfc9-81474d67d928","Type":"ContainerStarted","Data":"39ff5a2a98843182a0574523530ac8dc4f9f7a42fe1e1ed590366daa26ed16aa"} Dec 09 10:09:17 crc kubenswrapper[4824]: I1209 10:09:17.993980 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-z775c" Dec 09 10:09:17 crc kubenswrapper[4824]: I1209 10:09:17.995940 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" event={"ID":"9177fe98-0bd2-43b4-85bb-f9218d48211e","Type":"ContainerStarted","Data":"d2d1b5606cbde06cfba9a3a183389179cd58d0f9d03ccc9ee2a7f41baaa97c94"} Dec 09 10:09:17 crc kubenswrapper[4824]: I1209 10:09:17.996105 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" Dec 09 10:09:17 crc kubenswrapper[4824]: I1209 10:09:17.997868 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c2dd0680-d32f-45e8-b834-6a327d2eaa21","Type":"ContainerStarted","Data":"e2aa6bcd330d72dc81dd053337a4ea051d366f4865d00fa7ff413632d0470eb8"} Dec 09 10:09:17 crc kubenswrapper[4824]: I1209 10:09:17.998040 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 09 10:09:18 crc kubenswrapper[4824]: I1209 10:09:18.043797 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" podStartSLOduration=14.516071893 podStartE2EDuration="15.04375013s" podCreationTimestamp="2025-12-09 10:09:03 +0000 UTC" firstStartedPulling="2025-12-09 10:09:06.611126375 +0000 UTC m=+1302.945631042" lastFinishedPulling="2025-12-09 10:09:07.138804612 +0000 UTC m=+1303.473309279" observedRunningTime="2025-12-09 10:09:18.036301923 +0000 UTC m=+1314.370806600" watchObservedRunningTime="2025-12-09 10:09:18.04375013 +0000 UTC m=+1314.378254797" Dec 09 10:09:18 crc kubenswrapper[4824]: I1209 10:09:18.060131 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-z775c" podStartSLOduration=13.531016227 podStartE2EDuration="14.060107958s" podCreationTimestamp="2025-12-09 10:09:04 +0000 UTC" firstStartedPulling="2025-12-09 10:09:06.610936329 +0000 UTC m=+1302.945440996" lastFinishedPulling="2025-12-09 10:09:07.14002805 +0000 UTC m=+1303.474532727" observedRunningTime="2025-12-09 10:09:18.050677509 +0000 UTC m=+1314.385182176" watchObservedRunningTime="2025-12-09 10:09:18.060107958 +0000 UTC m=+1314.394612625" Dec 09 10:09:18 crc kubenswrapper[4824]: I1209 10:09:18.077916 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=11.484543938 podStartE2EDuration="14.077893512s" podCreationTimestamp="2025-12-09 10:09:04 +0000 UTC" firstStartedPulling="2025-12-09 10:09:06.968267756 +0000 UTC m=+1303.302772423" lastFinishedPulling="2025-12-09 10:09:09.56161732 +0000 UTC m=+1305.896121997" observedRunningTime="2025-12-09 10:09:18.067126571 +0000 UTC m=+1314.401631258" watchObservedRunningTime="2025-12-09 10:09:18.077893512 +0000 UTC m=+1314.412398179" Dec 09 10:09:18 crc kubenswrapper[4824]: I1209 10:09:18.408991 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 09 10:09:18 crc kubenswrapper[4824]: I1209 10:09:18.489945 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.115473 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-b025-account-create-update-cxmdd"] Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.117647 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b025-account-create-update-cxmdd" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.119612 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.126307 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-5vfks"] Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.127999 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-5vfks" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.135226 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b025-account-create-update-cxmdd"] Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.150721 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-5vfks"] Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.222741 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-ccjw2"] Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.224808 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ccjw2" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.234292 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-65e6-account-create-update-znnqb"] Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.236337 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-65e6-account-create-update-znnqb" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.248003 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-65e6-account-create-update-znnqb"] Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.248393 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.257950 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-ccjw2"] Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.264442 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/02d954cc-c1d4-45c8-86a2-fe001a222616-operator-scripts\") pod \"keystone-db-create-5vfks\" (UID: \"02d954cc-c1d4-45c8-86a2-fe001a222616\") " pod="openstack/keystone-db-create-5vfks" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.264548 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2d7v\" (UniqueName: \"kubernetes.io/projected/373c7c35-5998-40d4-b40b-5024d81e5567-kube-api-access-r2d7v\") pod \"keystone-b025-account-create-update-cxmdd\" (UID: \"373c7c35-5998-40d4-b40b-5024d81e5567\") " pod="openstack/keystone-b025-account-create-update-cxmdd" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.264768 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdl7p\" (UniqueName: \"kubernetes.io/projected/02d954cc-c1d4-45c8-86a2-fe001a222616-kube-api-access-mdl7p\") pod \"keystone-db-create-5vfks\" (UID: \"02d954cc-c1d4-45c8-86a2-fe001a222616\") " pod="openstack/keystone-db-create-5vfks" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.265122 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/373c7c35-5998-40d4-b40b-5024d81e5567-operator-scripts\") pod \"keystone-b025-account-create-update-cxmdd\" (UID: \"373c7c35-5998-40d4-b40b-5024d81e5567\") " pod="openstack/keystone-b025-account-create-update-cxmdd" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.371252 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/02d954cc-c1d4-45c8-86a2-fe001a222616-operator-scripts\") pod \"keystone-db-create-5vfks\" (UID: \"02d954cc-c1d4-45c8-86a2-fe001a222616\") " pod="openstack/keystone-db-create-5vfks" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.371615 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2d7v\" (UniqueName: \"kubernetes.io/projected/373c7c35-5998-40d4-b40b-5024d81e5567-kube-api-access-r2d7v\") pod \"keystone-b025-account-create-update-cxmdd\" (UID: \"373c7c35-5998-40d4-b40b-5024d81e5567\") " pod="openstack/keystone-b025-account-create-update-cxmdd" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.371717 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5p9f9\" (UniqueName: \"kubernetes.io/projected/1ec0f64d-691b-4433-a92b-ae6ab0ae4971-kube-api-access-5p9f9\") pod \"placement-65e6-account-create-update-znnqb\" (UID: \"1ec0f64d-691b-4433-a92b-ae6ab0ae4971\") " pod="openstack/placement-65e6-account-create-update-znnqb" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.371827 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdl7p\" (UniqueName: \"kubernetes.io/projected/02d954cc-c1d4-45c8-86a2-fe001a222616-kube-api-access-mdl7p\") pod \"keystone-db-create-5vfks\" (UID: \"02d954cc-c1d4-45c8-86a2-fe001a222616\") " pod="openstack/keystone-db-create-5vfks" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.372016 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/373c7c35-5998-40d4-b40b-5024d81e5567-operator-scripts\") pod \"keystone-b025-account-create-update-cxmdd\" (UID: \"373c7c35-5998-40d4-b40b-5024d81e5567\") " pod="openstack/keystone-b025-account-create-update-cxmdd" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.372171 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ec0f64d-691b-4433-a92b-ae6ab0ae4971-operator-scripts\") pod \"placement-65e6-account-create-update-znnqb\" (UID: \"1ec0f64d-691b-4433-a92b-ae6ab0ae4971\") " pod="openstack/placement-65e6-account-create-update-znnqb" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.372304 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/245fbd50-9e68-416a-a260-56363b243f55-operator-scripts\") pod \"placement-db-create-ccjw2\" (UID: \"245fbd50-9e68-416a-a260-56363b243f55\") " pod="openstack/placement-db-create-ccjw2" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.372389 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgm5w\" (UniqueName: \"kubernetes.io/projected/245fbd50-9e68-416a-a260-56363b243f55-kube-api-access-lgm5w\") pod \"placement-db-create-ccjw2\" (UID: \"245fbd50-9e68-416a-a260-56363b243f55\") " pod="openstack/placement-db-create-ccjw2" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.373182 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/02d954cc-c1d4-45c8-86a2-fe001a222616-operator-scripts\") pod \"keystone-db-create-5vfks\" (UID: \"02d954cc-c1d4-45c8-86a2-fe001a222616\") " pod="openstack/keystone-db-create-5vfks" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.374460 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/373c7c35-5998-40d4-b40b-5024d81e5567-operator-scripts\") pod \"keystone-b025-account-create-update-cxmdd\" (UID: \"373c7c35-5998-40d4-b40b-5024d81e5567\") " pod="openstack/keystone-b025-account-create-update-cxmdd" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.394335 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdl7p\" (UniqueName: \"kubernetes.io/projected/02d954cc-c1d4-45c8-86a2-fe001a222616-kube-api-access-mdl7p\") pod \"keystone-db-create-5vfks\" (UID: \"02d954cc-c1d4-45c8-86a2-fe001a222616\") " pod="openstack/keystone-db-create-5vfks" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.394398 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2d7v\" (UniqueName: \"kubernetes.io/projected/373c7c35-5998-40d4-b40b-5024d81e5567-kube-api-access-r2d7v\") pod \"keystone-b025-account-create-update-cxmdd\" (UID: \"373c7c35-5998-40d4-b40b-5024d81e5567\") " pod="openstack/keystone-b025-account-create-update-cxmdd" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.446817 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b025-account-create-update-cxmdd" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.458988 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-5vfks" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.474165 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ec0f64d-691b-4433-a92b-ae6ab0ae4971-operator-scripts\") pod \"placement-65e6-account-create-update-znnqb\" (UID: \"1ec0f64d-691b-4433-a92b-ae6ab0ae4971\") " pod="openstack/placement-65e6-account-create-update-znnqb" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.474238 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/245fbd50-9e68-416a-a260-56363b243f55-operator-scripts\") pod \"placement-db-create-ccjw2\" (UID: \"245fbd50-9e68-416a-a260-56363b243f55\") " pod="openstack/placement-db-create-ccjw2" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.474258 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgm5w\" (UniqueName: \"kubernetes.io/projected/245fbd50-9e68-416a-a260-56363b243f55-kube-api-access-lgm5w\") pod \"placement-db-create-ccjw2\" (UID: \"245fbd50-9e68-416a-a260-56363b243f55\") " pod="openstack/placement-db-create-ccjw2" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.474317 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5p9f9\" (UniqueName: \"kubernetes.io/projected/1ec0f64d-691b-4433-a92b-ae6ab0ae4971-kube-api-access-5p9f9\") pod \"placement-65e6-account-create-update-znnqb\" (UID: \"1ec0f64d-691b-4433-a92b-ae6ab0ae4971\") " pod="openstack/placement-65e6-account-create-update-znnqb" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.475420 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ec0f64d-691b-4433-a92b-ae6ab0ae4971-operator-scripts\") pod \"placement-65e6-account-create-update-znnqb\" (UID: \"1ec0f64d-691b-4433-a92b-ae6ab0ae4971\") " pod="openstack/placement-65e6-account-create-update-znnqb" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.476032 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/245fbd50-9e68-416a-a260-56363b243f55-operator-scripts\") pod \"placement-db-create-ccjw2\" (UID: \"245fbd50-9e68-416a-a260-56363b243f55\") " pod="openstack/placement-db-create-ccjw2" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.496181 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgm5w\" (UniqueName: \"kubernetes.io/projected/245fbd50-9e68-416a-a260-56363b243f55-kube-api-access-lgm5w\") pod \"placement-db-create-ccjw2\" (UID: \"245fbd50-9e68-416a-a260-56363b243f55\") " pod="openstack/placement-db-create-ccjw2" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.497601 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5p9f9\" (UniqueName: \"kubernetes.io/projected/1ec0f64d-691b-4433-a92b-ae6ab0ae4971-kube-api-access-5p9f9\") pod \"placement-65e6-account-create-update-znnqb\" (UID: \"1ec0f64d-691b-4433-a92b-ae6ab0ae4971\") " pod="openstack/placement-65e6-account-create-update-znnqb" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.553912 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ccjw2" Dec 09 10:09:19 crc kubenswrapper[4824]: I1209 10:09:19.569194 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-65e6-account-create-update-znnqb" Dec 09 10:09:21 crc kubenswrapper[4824]: I1209 10:09:21.705011 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-ntx24" podUID="ad8f524a-8c95-47ab-b74d-9f83331fce76" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 10:09:23 crc kubenswrapper[4824]: I1209 10:09:23.284252 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-5xfv2" podUID="87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 10:09:23 crc kubenswrapper[4824]: I1209 10:09:23.284372 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-5xfv2" podUID="87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 10:09:23 crc kubenswrapper[4824]: I1209 10:09:23.835765 4824 patch_prober.go:28] interesting pod/router-default-5444994796-mj7jx container/router namespace/openshift-ingress: Liveness probe status=failure output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 10:09:23 crc kubenswrapper[4824]: I1209 10:09:23.835956 4824 patch_prober.go:28] interesting pod/router-default-5444994796-mj7jx container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 10:09:23 crc kubenswrapper[4824]: I1209 10:09:23.836066 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 10:09:23 crc kubenswrapper[4824]: I1209 10:09:23.835955 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 10:09:24 crc kubenswrapper[4824]: I1209 10:09:24.690118 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-vz44w container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 10:09:24 crc kubenswrapper[4824]: I1209 10:09:24.690577 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" podUID="93c989b8-7b86-4339-bbab-5886c7d13dc9" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.55:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 10:09:24 crc kubenswrapper[4824]: I1209 10:09:24.851189 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-bvp6v container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 10:09:24 crc kubenswrapper[4824]: I1209 10:09:24.851262 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" podUID="858d899b-800f-4639-8fbd-4f1ccad44991" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 10:09:29 crc kubenswrapper[4824]: I1209 10:09:29.689907 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-vz44w container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 10:09:29 crc kubenswrapper[4824]: I1209 10:09:29.690019 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" podUID="93c989b8-7b86-4339-bbab-5886c7d13dc9" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.55:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 10:09:29 crc kubenswrapper[4824]: I1209 10:09:29.852505 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-bvp6v container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 10:09:29 crc kubenswrapper[4824]: I1209 10:09:29.852649 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" podUID="858d899b-800f-4639-8fbd-4f1ccad44991" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 10:09:31 crc kubenswrapper[4824]: I1209 10:09:31.706192 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-ntx24" podUID="ad8f524a-8c95-47ab-b74d-9f83331fce76" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 10:09:33 crc kubenswrapper[4824]: I1209 10:09:33.284018 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-5xfv2" podUID="87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 10:09:33 crc kubenswrapper[4824]: I1209 10:09:33.284178 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-5xfv2" podUID="87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 10:09:33 crc kubenswrapper[4824]: I1209 10:09:33.835543 4824 patch_prober.go:28] interesting pod/router-default-5444994796-mj7jx container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 10:09:33 crc kubenswrapper[4824]: I1209 10:09:33.835669 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 10:09:33 crc kubenswrapper[4824]: I1209 10:09:33.836036 4824 patch_prober.go:28] interesting pod/router-default-5444994796-mj7jx container/router namespace/openshift-ingress: Liveness probe status=failure output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 10:09:33 crc kubenswrapper[4824]: I1209 10:09:33.836139 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.282127 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-77dcf56988-s8mzf" podUID="721e1031-d80c-406a-ad29-17879d58c3c3" containerName="console" containerID="cri-o://88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857" gracePeriod=11 Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.283912 4824 trace.go:236] Trace[1027488336]: "Calculate volume metrics of mysql-db for pod openstack/openstack-galera-0" (09-Dec-2025 10:09:31.505) (total time: 2778ms): Dec 09 10:09:34 crc kubenswrapper[4824]: Trace[1027488336]: [2.778261287s] [2.778261287s] END Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.283977 4824 trace.go:236] Trace[1762339050]: "Calculate volume metrics of persistence for pod openstack/rabbitmq-server-1" (09-Dec-2025 10:09:32.523) (total time: 1760ms): Dec 09 10:09:34 crc kubenswrapper[4824]: Trace[1762339050]: [1.760416062s] [1.760416062s] END Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.286337 4824 trace.go:236] Trace[1584695829]: "Calculate volume metrics of ovndbcluster-sb-etc-ovn for pod openstack/ovsdbserver-sb-0" (09-Dec-2025 10:09:23.904) (total time: 10381ms): Dec 09 10:09:34 crc kubenswrapper[4824]: Trace[1584695829]: [10.381895089s] [10.381895089s] END Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.287489 4824 patch_prober.go:28] interesting pod/etcd-crc container/etcd namespace/openshift-etcd: Liveness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=failed to establish etcd client: giving up getting a cached client after 3 tries Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.287528 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-etcd/etcd-crc" podUID="2139d3e2895fc6797b9c76a1b4c9886d" containerName="etcd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 09 10:09:34 crc kubenswrapper[4824]: E1209 10:09:34.371466 4824 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="14.462s" Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.371846 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.371884 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-z775c" Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.495637 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-fmfsh"] Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.508743 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-fmfsh"] Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.508804 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-7029-account-create-update-dkjjt"] Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.515741 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-7029-account-create-update-dkjjt"] Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.515794 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-p4czz"] Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.515816 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-t4gz7"] Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.523487 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-t4gz7"] Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.523530 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.524765 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-7029-account-create-update-dkjjt" Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.586962 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-db-secret" Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.648612 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-t4gz7" Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.654840 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-fmfsh" Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.930817 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.959664 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-pztmz"] Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.961561 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-pztmz"] Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.961637 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.961674 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pztmz" Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.961722 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-a94f-account-create-update-f2dbj"] Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.964858 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-a94f-account-create-update-f2dbj"] Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.964898 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-ncg6n"] Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.965905 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-ncg6n"] Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.965931 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-ncg6n"] Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.965949 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-cf5wn"] Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.967060 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a94f-account-create-update-f2dbj" Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.975762 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.976345 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-ncg6n" Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.981585 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-lcbb7" Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.981806 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.981922 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.982123 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.991743 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-cf5wn"] Dec 09 10:09:34 crc kubenswrapper[4824]: I1209 10:09:34.992196 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.002149 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.006625 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.014508 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82ce77fc-cdbd-4df3-b067-195040a2836e-dns-svc\") pod \"dnsmasq-dns-698758b865-t4gz7\" (UID: \"82ce77fc-cdbd-4df3-b067-195040a2836e\") " pod="openstack/dnsmasq-dns-698758b865-t4gz7" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.014566 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82ce77fc-cdbd-4df3-b067-195040a2836e-config\") pod \"dnsmasq-dns-698758b865-t4gz7\" (UID: \"82ce77fc-cdbd-4df3-b067-195040a2836e\") " pod="openstack/dnsmasq-dns-698758b865-t4gz7" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.014597 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83cb8977-7594-4739-90ae-ba915822102e-operator-scripts\") pod \"mysqld-exporter-7029-account-create-update-dkjjt\" (UID: \"83cb8977-7594-4739-90ae-ba915822102e\") " pod="openstack/mysqld-exporter-7029-account-create-update-dkjjt" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.014640 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq58n\" (UniqueName: \"kubernetes.io/projected/033734aa-97c9-498f-8bae-992507b79f71-kube-api-access-rq58n\") pod \"mysqld-exporter-openstack-db-create-fmfsh\" (UID: \"033734aa-97c9-498f-8bae-992507b79f71\") " pod="openstack/mysqld-exporter-openstack-db-create-fmfsh" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.014678 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82ce77fc-cdbd-4df3-b067-195040a2836e-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-t4gz7\" (UID: \"82ce77fc-cdbd-4df3-b067-195040a2836e\") " pod="openstack/dnsmasq-dns-698758b865-t4gz7" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.016303 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82ce77fc-cdbd-4df3-b067-195040a2836e-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-t4gz7\" (UID: \"82ce77fc-cdbd-4df3-b067-195040a2836e\") " pod="openstack/dnsmasq-dns-698758b865-t4gz7" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.016388 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfcbm\" (UniqueName: \"kubernetes.io/projected/83cb8977-7594-4739-90ae-ba915822102e-kube-api-access-cfcbm\") pod \"mysqld-exporter-7029-account-create-update-dkjjt\" (UID: \"83cb8977-7594-4739-90ae-ba915822102e\") " pod="openstack/mysqld-exporter-7029-account-create-update-dkjjt" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.016562 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/033734aa-97c9-498f-8bae-992507b79f71-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-fmfsh\" (UID: \"033734aa-97c9-498f-8bae-992507b79f71\") " pod="openstack/mysqld-exporter-openstack-db-create-fmfsh" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.016825 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbnvd\" (UniqueName: \"kubernetes.io/projected/82ce77fc-cdbd-4df3-b067-195040a2836e-kube-api-access-rbnvd\") pod \"dnsmasq-dns-698758b865-t4gz7\" (UID: \"82ce77fc-cdbd-4df3-b067-195040a2836e\") " pod="openstack/dnsmasq-dns-698758b865-t4gz7" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.017081 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.119853 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.129079 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-dispersionconf\") pod \"swift-ring-rebalance-cf5wn\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.131400 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-scripts\") pod \"swift-ring-rebalance-cf5wn\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.131501 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-cache\") pod \"swift-storage-0\" (UID: \"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f\") " pod="openstack/swift-storage-0" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.131525 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h85hd\" (UniqueName: \"kubernetes.io/projected/1c6345d1-110d-4671-b27b-f8d2abcf4d4f-kube-api-access-h85hd\") pod \"glance-a94f-account-create-update-f2dbj\" (UID: \"1c6345d1-110d-4671-b27b-f8d2abcf4d4f\") " pod="openstack/glance-a94f-account-create-update-f2dbj" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.131667 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxp2s\" (UniqueName: \"kubernetes.io/projected/7b6a6023-0aa7-4778-be37-6de2222088c4-kube-api-access-xxp2s\") pod \"glance-db-create-pztmz\" (UID: \"7b6a6023-0aa7-4778-be37-6de2222088c4\") " pod="openstack/glance-db-create-pztmz" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.131694 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7bks\" (UniqueName: \"kubernetes.io/projected/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-kube-api-access-w7bks\") pod \"swift-ring-rebalance-cf5wn\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.131734 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82ce77fc-cdbd-4df3-b067-195040a2836e-dns-svc\") pod \"dnsmasq-dns-698758b865-t4gz7\" (UID: \"82ce77fc-cdbd-4df3-b067-195040a2836e\") " pod="openstack/dnsmasq-dns-698758b865-t4gz7" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.131795 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82ce77fc-cdbd-4df3-b067-195040a2836e-config\") pod \"dnsmasq-dns-698758b865-t4gz7\" (UID: \"82ce77fc-cdbd-4df3-b067-195040a2836e\") " pod="openstack/dnsmasq-dns-698758b865-t4gz7" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.131828 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83cb8977-7594-4739-90ae-ba915822102e-operator-scripts\") pod \"mysqld-exporter-7029-account-create-update-dkjjt\" (UID: \"83cb8977-7594-4739-90ae-ba915822102e\") " pod="openstack/mysqld-exporter-7029-account-create-update-dkjjt" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.131884 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq58n\" (UniqueName: \"kubernetes.io/projected/033734aa-97c9-498f-8bae-992507b79f71-kube-api-access-rq58n\") pod \"mysqld-exporter-openstack-db-create-fmfsh\" (UID: \"033734aa-97c9-498f-8bae-992507b79f71\") " pod="openstack/mysqld-exporter-openstack-db-create-fmfsh" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.131909 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-swiftconf\") pod \"swift-ring-rebalance-cf5wn\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.131952 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82ce77fc-cdbd-4df3-b067-195040a2836e-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-t4gz7\" (UID: \"82ce77fc-cdbd-4df3-b067-195040a2836e\") " pod="openstack/dnsmasq-dns-698758b865-t4gz7" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.131973 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-lock\") pod \"swift-storage-0\" (UID: \"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f\") " pod="openstack/swift-storage-0" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.132043 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-etc-swift\") pod \"swift-storage-0\" (UID: \"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f\") " pod="openstack/swift-storage-0" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.132077 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-ring-data-devices\") pod \"swift-ring-rebalance-cf5wn\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.132111 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-combined-ca-bundle\") pod \"swift-ring-rebalance-cf5wn\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.132168 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftsjr\" (UniqueName: \"kubernetes.io/projected/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-kube-api-access-ftsjr\") pod \"swift-storage-0\" (UID: \"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f\") " pod="openstack/swift-storage-0" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.132220 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ac95c8f5-28a6-414e-b2e7-db94b729c44f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ac95c8f5-28a6-414e-b2e7-db94b729c44f\") pod \"swift-storage-0\" (UID: \"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f\") " pod="openstack/swift-storage-0" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.132268 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82ce77fc-cdbd-4df3-b067-195040a2836e-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-t4gz7\" (UID: \"82ce77fc-cdbd-4df3-b067-195040a2836e\") " pod="openstack/dnsmasq-dns-698758b865-t4gz7" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.132319 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfcbm\" (UniqueName: \"kubernetes.io/projected/83cb8977-7594-4739-90ae-ba915822102e-kube-api-access-cfcbm\") pod \"mysqld-exporter-7029-account-create-update-dkjjt\" (UID: \"83cb8977-7594-4739-90ae-ba915822102e\") " pod="openstack/mysqld-exporter-7029-account-create-update-dkjjt" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.132372 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c6345d1-110d-4671-b27b-f8d2abcf4d4f-operator-scripts\") pod \"glance-a94f-account-create-update-f2dbj\" (UID: \"1c6345d1-110d-4671-b27b-f8d2abcf4d4f\") " pod="openstack/glance-a94f-account-create-update-f2dbj" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.133561 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b6a6023-0aa7-4778-be37-6de2222088c4-operator-scripts\") pod \"glance-db-create-pztmz\" (UID: \"7b6a6023-0aa7-4778-be37-6de2222088c4\") " pod="openstack/glance-db-create-pztmz" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.133596 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-etc-swift\") pod \"swift-ring-rebalance-cf5wn\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.133629 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/033734aa-97c9-498f-8bae-992507b79f71-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-fmfsh\" (UID: \"033734aa-97c9-498f-8bae-992507b79f71\") " pod="openstack/mysqld-exporter-openstack-db-create-fmfsh" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.133717 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbnvd\" (UniqueName: \"kubernetes.io/projected/82ce77fc-cdbd-4df3-b067-195040a2836e-kube-api-access-rbnvd\") pod \"dnsmasq-dns-698758b865-t4gz7\" (UID: \"82ce77fc-cdbd-4df3-b067-195040a2836e\") " pod="openstack/dnsmasq-dns-698758b865-t4gz7" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.135992 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82ce77fc-cdbd-4df3-b067-195040a2836e-dns-svc\") pod \"dnsmasq-dns-698758b865-t4gz7\" (UID: \"82ce77fc-cdbd-4df3-b067-195040a2836e\") " pod="openstack/dnsmasq-dns-698758b865-t4gz7" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.137146 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82ce77fc-cdbd-4df3-b067-195040a2836e-config\") pod \"dnsmasq-dns-698758b865-t4gz7\" (UID: \"82ce77fc-cdbd-4df3-b067-195040a2836e\") " pod="openstack/dnsmasq-dns-698758b865-t4gz7" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.138250 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83cb8977-7594-4739-90ae-ba915822102e-operator-scripts\") pod \"mysqld-exporter-7029-account-create-update-dkjjt\" (UID: \"83cb8977-7594-4739-90ae-ba915822102e\") " pod="openstack/mysqld-exporter-7029-account-create-update-dkjjt" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.139062 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82ce77fc-cdbd-4df3-b067-195040a2836e-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-t4gz7\" (UID: \"82ce77fc-cdbd-4df3-b067-195040a2836e\") " pod="openstack/dnsmasq-dns-698758b865-t4gz7" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.139147 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82ce77fc-cdbd-4df3-b067-195040a2836e-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-t4gz7\" (UID: \"82ce77fc-cdbd-4df3-b067-195040a2836e\") " pod="openstack/dnsmasq-dns-698758b865-t4gz7" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.140389 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/033734aa-97c9-498f-8bae-992507b79f71-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-fmfsh\" (UID: \"033734aa-97c9-498f-8bae-992507b79f71\") " pod="openstack/mysqld-exporter-openstack-db-create-fmfsh" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.175022 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbnvd\" (UniqueName: \"kubernetes.io/projected/82ce77fc-cdbd-4df3-b067-195040a2836e-kube-api-access-rbnvd\") pod \"dnsmasq-dns-698758b865-t4gz7\" (UID: \"82ce77fc-cdbd-4df3-b067-195040a2836e\") " pod="openstack/dnsmasq-dns-698758b865-t4gz7" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.176989 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfcbm\" (UniqueName: \"kubernetes.io/projected/83cb8977-7594-4739-90ae-ba915822102e-kube-api-access-cfcbm\") pod \"mysqld-exporter-7029-account-create-update-dkjjt\" (UID: \"83cb8977-7594-4739-90ae-ba915822102e\") " pod="openstack/mysqld-exporter-7029-account-create-update-dkjjt" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.181175 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq58n\" (UniqueName: \"kubernetes.io/projected/033734aa-97c9-498f-8bae-992507b79f71-kube-api-access-rq58n\") pod \"mysqld-exporter-openstack-db-create-fmfsh\" (UID: \"033734aa-97c9-498f-8bae-992507b79f71\") " pod="openstack/mysqld-exporter-openstack-db-create-fmfsh" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.225333 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"30eb761a-7bff-43a5-adfb-40e90e17e4a6","Type":"ContainerStarted","Data":"24701089b3049252c9f2d6a4217bcb6c2856b5c88e9859477a2fe57787dc088a"} Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.231164 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"cee7a7d4-090b-4db0-abf1-d3f861f8e594","Type":"ContainerStarted","Data":"9340d5745e62248e475ec029542c6a7717fa8bb811727daef75c5eb26017e816"} Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.235852 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-swiftconf\") pod \"swift-ring-rebalance-cf5wn\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.236613 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-lock\") pod \"swift-storage-0\" (UID: \"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f\") " pod="openstack/swift-storage-0" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.236833 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-etc-swift\") pod \"swift-storage-0\" (UID: \"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f\") " pod="openstack/swift-storage-0" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.236935 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-ring-data-devices\") pod \"swift-ring-rebalance-cf5wn\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.237057 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-combined-ca-bundle\") pod \"swift-ring-rebalance-cf5wn\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.237211 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftsjr\" (UniqueName: \"kubernetes.io/projected/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-kube-api-access-ftsjr\") pod \"swift-storage-0\" (UID: \"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f\") " pod="openstack/swift-storage-0" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.237328 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ac95c8f5-28a6-414e-b2e7-db94b729c44f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ac95c8f5-28a6-414e-b2e7-db94b729c44f\") pod \"swift-storage-0\" (UID: \"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f\") " pod="openstack/swift-storage-0" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.237476 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c6345d1-110d-4671-b27b-f8d2abcf4d4f-operator-scripts\") pod \"glance-a94f-account-create-update-f2dbj\" (UID: \"1c6345d1-110d-4671-b27b-f8d2abcf4d4f\") " pod="openstack/glance-a94f-account-create-update-f2dbj" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.237562 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b6a6023-0aa7-4778-be37-6de2222088c4-operator-scripts\") pod \"glance-db-create-pztmz\" (UID: \"7b6a6023-0aa7-4778-be37-6de2222088c4\") " pod="openstack/glance-db-create-pztmz" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.237646 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-etc-swift\") pod \"swift-ring-rebalance-cf5wn\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.237893 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-dispersionconf\") pod \"swift-ring-rebalance-cf5wn\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.238294 4824 generic.go:334] "Generic (PLEG): container finished" podID="8edb90f3-f84b-4700-993f-21ffa950b620" containerID="7cf4f49e8f8e0800f59be8f22816ba55ee3205fc0e8f6c81552552f0e44a23f7" exitCode=0 Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.238413 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8edb90f3-f84b-4700-993f-21ffa950b620","Type":"ContainerDied","Data":"7cf4f49e8f8e0800f59be8f22816ba55ee3205fc0e8f6c81552552f0e44a23f7"} Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.238583 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-scripts\") pod \"swift-ring-rebalance-cf5wn\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.239194 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-cache\") pod \"swift-storage-0\" (UID: \"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f\") " pod="openstack/swift-storage-0" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.239270 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h85hd\" (UniqueName: \"kubernetes.io/projected/1c6345d1-110d-4671-b27b-f8d2abcf4d4f-kube-api-access-h85hd\") pod \"glance-a94f-account-create-update-f2dbj\" (UID: \"1c6345d1-110d-4671-b27b-f8d2abcf4d4f\") " pod="openstack/glance-a94f-account-create-update-f2dbj" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.239402 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxp2s\" (UniqueName: \"kubernetes.io/projected/7b6a6023-0aa7-4778-be37-6de2222088c4-kube-api-access-xxp2s\") pod \"glance-db-create-pztmz\" (UID: \"7b6a6023-0aa7-4778-be37-6de2222088c4\") " pod="openstack/glance-db-create-pztmz" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.239596 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7bks\" (UniqueName: \"kubernetes.io/projected/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-kube-api-access-w7bks\") pod \"swift-ring-rebalance-cf5wn\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.240701 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c6345d1-110d-4671-b27b-f8d2abcf4d4f-operator-scripts\") pod \"glance-a94f-account-create-update-f2dbj\" (UID: \"1c6345d1-110d-4671-b27b-f8d2abcf4d4f\") " pod="openstack/glance-a94f-account-create-update-f2dbj" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.242200 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b6a6023-0aa7-4778-be37-6de2222088c4-operator-scripts\") pod \"glance-db-create-pztmz\" (UID: \"7b6a6023-0aa7-4778-be37-6de2222088c4\") " pod="openstack/glance-db-create-pztmz" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.242825 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-scripts\") pod \"swift-ring-rebalance-cf5wn\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.243271 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-etc-swift\") pod \"swift-ring-rebalance-cf5wn\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.243383 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-ring-data-devices\") pod \"swift-ring-rebalance-cf5wn\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.243901 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-lock\") pod \"swift-storage-0\" (UID: \"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f\") " pod="openstack/swift-storage-0" Dec 09 10:09:35 crc kubenswrapper[4824]: E1209 10:09:35.244005 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 09 10:09:35 crc kubenswrapper[4824]: E1209 10:09:35.253517 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 09 10:09:35 crc kubenswrapper[4824]: E1209 10:09:35.253678 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-etc-swift podName:66fd04f1-a0d9-4295-ac42-e5b20acc7c1f nodeName:}" failed. No retries permitted until 2025-12-09 10:09:35.753644274 +0000 UTC m=+1332.088148941 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-etc-swift") pod "swift-storage-0" (UID: "66fd04f1-a0d9-4295-ac42-e5b20acc7c1f") : configmap "swift-ring-files" not found Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.244404 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-cache\") pod \"swift-storage-0\" (UID: \"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f\") " pod="openstack/swift-storage-0" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.254190 4824 generic.go:334] "Generic (PLEG): container finished" podID="07c1822c-1257-4d8c-9a17-d921929000fa" containerID="b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c" exitCode=0 Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.254260 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"07c1822c-1257-4d8c-9a17-d921929000fa","Type":"ContainerDied","Data":"b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c"} Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.255525 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.255553 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ac95c8f5-28a6-414e-b2e7-db94b729c44f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ac95c8f5-28a6-414e-b2e7-db94b729c44f\") pod \"swift-storage-0\" (UID: \"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e35374844ad6b1c170ad072eb6d5c1582c28abb7c943076f44f5fd9b9690927d/globalmount\"" pod="openstack/swift-storage-0" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.263496 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-77dcf56988-s8mzf_721e1031-d80c-406a-ad29-17879d58c3c3/console/0.log" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.263594 4824 generic.go:334] "Generic (PLEG): container finished" podID="721e1031-d80c-406a-ad29-17879d58c3c3" containerID="88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857" exitCode=2 Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.263941 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" podUID="9177fe98-0bd2-43b4-85bb-f9218d48211e" containerName="dnsmasq-dns" containerID="cri-o://d2d1b5606cbde06cfba9a3a183389179cd58d0f9d03ccc9ee2a7f41baaa97c94" gracePeriod=10 Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.264101 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-77dcf56988-s8mzf" event={"ID":"721e1031-d80c-406a-ad29-17879d58c3c3","Type":"ContainerDied","Data":"88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857"} Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.281412 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-swiftconf\") pod \"swift-ring-rebalance-cf5wn\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.285511 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-combined-ca-bundle\") pod \"swift-ring-rebalance-cf5wn\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.300301 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftsjr\" (UniqueName: \"kubernetes.io/projected/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-kube-api-access-ftsjr\") pod \"swift-storage-0\" (UID: \"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f\") " pod="openstack/swift-storage-0" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.303056 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-dispersionconf\") pod \"swift-ring-rebalance-cf5wn\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.306362 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxp2s\" (UniqueName: \"kubernetes.io/projected/7b6a6023-0aa7-4778-be37-6de2222088c4-kube-api-access-xxp2s\") pod \"glance-db-create-pztmz\" (UID: \"7b6a6023-0aa7-4778-be37-6de2222088c4\") " pod="openstack/glance-db-create-pztmz" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.310931 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7bks\" (UniqueName: \"kubernetes.io/projected/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-kube-api-access-w7bks\") pod \"swift-ring-rebalance-cf5wn\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.312502 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h85hd\" (UniqueName: \"kubernetes.io/projected/1c6345d1-110d-4671-b27b-f8d2abcf4d4f-kube-api-access-h85hd\") pod \"glance-a94f-account-create-update-f2dbj\" (UID: \"1c6345d1-110d-4671-b27b-f8d2abcf4d4f\") " pod="openstack/glance-a94f-account-create-update-f2dbj" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.401464 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-985bt" podUID="961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d" containerName="ovn-controller" probeResult="failure" output=< Dec 09 10:09:35 crc kubenswrapper[4824]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 09 10:09:35 crc kubenswrapper[4824]: > Dec 09 10:09:35 crc kubenswrapper[4824]: E1209 10:09:35.409567 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c1822c_1257_4d8c_9a17_d921929000fa.slice/crio-b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c1822c_1257_4d8c_9a17_d921929000fa.slice/crio-conmon-b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8edb90f3_f84b_4700_993f_21ffa950b620.slice/crio-7cf4f49e8f8e0800f59be8f22816ba55ee3205fc0e8f6c81552552f0e44a23f7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721e1031_d80c_406a_ad29_17879d58c3c3.slice/crio-88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857.scope\": RecentStats: unable to find data in memory cache]" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.426598 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-rl4nw" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.442422 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-rl4nw" Dec 09 10:09:35 crc kubenswrapper[4824]: E1209 10:09:35.447216 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721e1031_d80c_406a_ad29_17879d58c3c3.slice/crio-88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721e1031_d80c_406a_ad29_17879d58c3c3.slice/crio-conmon-88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c1822c_1257_4d8c_9a17_d921929000fa.slice/crio-b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c.scope\": RecentStats: unable to find data in memory cache]" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.500998 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ac95c8f5-28a6-414e-b2e7-db94b729c44f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ac95c8f5-28a6-414e-b2e7-db94b729c44f\") pod \"swift-storage-0\" (UID: \"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f\") " pod="openstack/swift-storage-0" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.600607 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-985bt" podUID="961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d" containerName="ovn-controller" probeResult="failure" output=< Dec 09 10:09:35 crc kubenswrapper[4824]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 09 10:09:35 crc kubenswrapper[4824]: > Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.744008 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-7029-account-create-update-dkjjt" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.758023 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-t4gz7" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.765036 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-985bt-config-x2qzv"] Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.766840 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-985bt-config-x2qzv" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.769631 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.779360 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-fmfsh" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.795083 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-etc-swift\") pod \"swift-storage-0\" (UID: \"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f\") " pod="openstack/swift-storage-0" Dec 09 10:09:35 crc kubenswrapper[4824]: E1209 10:09:35.796088 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 09 10:09:35 crc kubenswrapper[4824]: E1209 10:09:35.796104 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 09 10:09:35 crc kubenswrapper[4824]: E1209 10:09:35.796143 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-etc-swift podName:66fd04f1-a0d9-4295-ac42-e5b20acc7c1f nodeName:}" failed. No retries permitted until 2025-12-09 10:09:36.796127111 +0000 UTC m=+1333.130631778 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-etc-swift") pod "swift-storage-0" (UID: "66fd04f1-a0d9-4295-ac42-e5b20acc7c1f") : configmap "swift-ring-files" not found Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.800130 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a94f-account-create-update-f2dbj" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.831693 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-985bt-config-x2qzv"] Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.867376 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b025-account-create-update-cxmdd"] Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.897990 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2vhw\" (UniqueName: \"kubernetes.io/projected/79e597f5-64c0-4250-b3d4-69ac3db9cb69-kube-api-access-f2vhw\") pod \"ovn-controller-985bt-config-x2qzv\" (UID: \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\") " pod="openstack/ovn-controller-985bt-config-x2qzv" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.898285 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/79e597f5-64c0-4250-b3d4-69ac3db9cb69-var-run\") pod \"ovn-controller-985bt-config-x2qzv\" (UID: \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\") " pod="openstack/ovn-controller-985bt-config-x2qzv" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.898351 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79e597f5-64c0-4250-b3d4-69ac3db9cb69-scripts\") pod \"ovn-controller-985bt-config-x2qzv\" (UID: \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\") " pod="openstack/ovn-controller-985bt-config-x2qzv" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.898399 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/79e597f5-64c0-4250-b3d4-69ac3db9cb69-additional-scripts\") pod \"ovn-controller-985bt-config-x2qzv\" (UID: \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\") " pod="openstack/ovn-controller-985bt-config-x2qzv" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.898422 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/79e597f5-64c0-4250-b3d4-69ac3db9cb69-var-log-ovn\") pod \"ovn-controller-985bt-config-x2qzv\" (UID: \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\") " pod="openstack/ovn-controller-985bt-config-x2qzv" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.898459 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/79e597f5-64c0-4250-b3d4-69ac3db9cb69-var-run-ovn\") pod \"ovn-controller-985bt-config-x2qzv\" (UID: \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\") " pod="openstack/ovn-controller-985bt-config-x2qzv" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.930920 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-ncg6n" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.972231 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.978976 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 09 10:09:35 crc kubenswrapper[4824]: I1209 10:09:35.994570 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pztmz" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.002684 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2vhw\" (UniqueName: \"kubernetes.io/projected/79e597f5-64c0-4250-b3d4-69ac3db9cb69-kube-api-access-f2vhw\") pod \"ovn-controller-985bt-config-x2qzv\" (UID: \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\") " pod="openstack/ovn-controller-985bt-config-x2qzv" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.002772 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/79e597f5-64c0-4250-b3d4-69ac3db9cb69-var-run\") pod \"ovn-controller-985bt-config-x2qzv\" (UID: \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\") " pod="openstack/ovn-controller-985bt-config-x2qzv" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.002839 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79e597f5-64c0-4250-b3d4-69ac3db9cb69-scripts\") pod \"ovn-controller-985bt-config-x2qzv\" (UID: \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\") " pod="openstack/ovn-controller-985bt-config-x2qzv" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.002895 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/79e597f5-64c0-4250-b3d4-69ac3db9cb69-additional-scripts\") pod \"ovn-controller-985bt-config-x2qzv\" (UID: \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\") " pod="openstack/ovn-controller-985bt-config-x2qzv" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.002944 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/79e597f5-64c0-4250-b3d4-69ac3db9cb69-var-log-ovn\") pod \"ovn-controller-985bt-config-x2qzv\" (UID: \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\") " pod="openstack/ovn-controller-985bt-config-x2qzv" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.003043 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/79e597f5-64c0-4250-b3d4-69ac3db9cb69-var-run-ovn\") pod \"ovn-controller-985bt-config-x2qzv\" (UID: \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\") " pod="openstack/ovn-controller-985bt-config-x2qzv" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.004685 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/79e597f5-64c0-4250-b3d4-69ac3db9cb69-var-run-ovn\") pod \"ovn-controller-985bt-config-x2qzv\" (UID: \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\") " pod="openstack/ovn-controller-985bt-config-x2qzv" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.004834 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/79e597f5-64c0-4250-b3d4-69ac3db9cb69-var-run\") pod \"ovn-controller-985bt-config-x2qzv\" (UID: \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\") " pod="openstack/ovn-controller-985bt-config-x2qzv" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.005364 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/79e597f5-64c0-4250-b3d4-69ac3db9cb69-var-log-ovn\") pod \"ovn-controller-985bt-config-x2qzv\" (UID: \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\") " pod="openstack/ovn-controller-985bt-config-x2qzv" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.007553 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79e597f5-64c0-4250-b3d4-69ac3db9cb69-scripts\") pod \"ovn-controller-985bt-config-x2qzv\" (UID: \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\") " pod="openstack/ovn-controller-985bt-config-x2qzv" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.008560 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/79e597f5-64c0-4250-b3d4-69ac3db9cb69-additional-scripts\") pod \"ovn-controller-985bt-config-x2qzv\" (UID: \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\") " pod="openstack/ovn-controller-985bt-config-x2qzv" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.041140 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2vhw\" (UniqueName: \"kubernetes.io/projected/79e597f5-64c0-4250-b3d4-69ac3db9cb69-kube-api-access-f2vhw\") pod \"ovn-controller-985bt-config-x2qzv\" (UID: \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\") " pod="openstack/ovn-controller-985bt-config-x2qzv" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.055077 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-77dcf56988-s8mzf_721e1031-d80c-406a-ad29-17879d58c3c3/console/0.log" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.055154 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.108767 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/721e1031-d80c-406a-ad29-17879d58c3c3-oauth-serving-cert\") pod \"721e1031-d80c-406a-ad29-17879d58c3c3\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.108975 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/721e1031-d80c-406a-ad29-17879d58c3c3-console-config\") pod \"721e1031-d80c-406a-ad29-17879d58c3c3\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.109011 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/721e1031-d80c-406a-ad29-17879d58c3c3-console-oauth-config\") pod \"721e1031-d80c-406a-ad29-17879d58c3c3\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.109064 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/721e1031-d80c-406a-ad29-17879d58c3c3-trusted-ca-bundle\") pod \"721e1031-d80c-406a-ad29-17879d58c3c3\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.109089 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/721e1031-d80c-406a-ad29-17879d58c3c3-service-ca\") pod \"721e1031-d80c-406a-ad29-17879d58c3c3\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.109135 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/721e1031-d80c-406a-ad29-17879d58c3c3-console-serving-cert\") pod \"721e1031-d80c-406a-ad29-17879d58c3c3\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.109229 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gj46v\" (UniqueName: \"kubernetes.io/projected/721e1031-d80c-406a-ad29-17879d58c3c3-kube-api-access-gj46v\") pod \"721e1031-d80c-406a-ad29-17879d58c3c3\" (UID: \"721e1031-d80c-406a-ad29-17879d58c3c3\") " Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.112148 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/721e1031-d80c-406a-ad29-17879d58c3c3-console-config" (OuterVolumeSpecName: "console-config") pod "721e1031-d80c-406a-ad29-17879d58c3c3" (UID: "721e1031-d80c-406a-ad29-17879d58c3c3"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.113676 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/721e1031-d80c-406a-ad29-17879d58c3c3-service-ca" (OuterVolumeSpecName: "service-ca") pod "721e1031-d80c-406a-ad29-17879d58c3c3" (UID: "721e1031-d80c-406a-ad29-17879d58c3c3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.114010 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/721e1031-d80c-406a-ad29-17879d58c3c3-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "721e1031-d80c-406a-ad29-17879d58c3c3" (UID: "721e1031-d80c-406a-ad29-17879d58c3c3"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.114208 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/721e1031-d80c-406a-ad29-17879d58c3c3-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "721e1031-d80c-406a-ad29-17879d58c3c3" (UID: "721e1031-d80c-406a-ad29-17879d58c3c3"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.117398 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/721e1031-d80c-406a-ad29-17879d58c3c3-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "721e1031-d80c-406a-ad29-17879d58c3c3" (UID: "721e1031-d80c-406a-ad29-17879d58c3c3"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.117422 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/721e1031-d80c-406a-ad29-17879d58c3c3-kube-api-access-gj46v" (OuterVolumeSpecName: "kube-api-access-gj46v") pod "721e1031-d80c-406a-ad29-17879d58c3c3" (UID: "721e1031-d80c-406a-ad29-17879d58c3c3"). InnerVolumeSpecName "kube-api-access-gj46v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.136202 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/721e1031-d80c-406a-ad29-17879d58c3c3-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "721e1031-d80c-406a-ad29-17879d58c3c3" (UID: "721e1031-d80c-406a-ad29-17879d58c3c3"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.192047 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.211851 4824 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/721e1031-d80c-406a-ad29-17879d58c3c3-console-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.211875 4824 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/721e1031-d80c-406a-ad29-17879d58c3c3-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.211884 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/721e1031-d80c-406a-ad29-17879d58c3c3-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.211892 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/721e1031-d80c-406a-ad29-17879d58c3c3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.211901 4824 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/721e1031-d80c-406a-ad29-17879d58c3c3-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.211912 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gj46v\" (UniqueName: \"kubernetes.io/projected/721e1031-d80c-406a-ad29-17879d58c3c3-kube-api-access-gj46v\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.211926 4824 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/721e1031-d80c-406a-ad29-17879d58c3c3-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.248107 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-ccjw2"] Dec 09 10:09:36 crc kubenswrapper[4824]: W1209 10:09:36.249167 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod245fbd50_9e68_416a_a260_56363b243f55.slice/crio-d962a2c4ff17cdd9b661c6b3df351c93b598c1a1f3a4a4c795c05ee815a5da08 WatchSource:0}: Error finding container d962a2c4ff17cdd9b661c6b3df351c93b598c1a1f3a4a4c795c05ee815a5da08: Status 404 returned error can't find the container with id d962a2c4ff17cdd9b661c6b3df351c93b598c1a1f3a4a4c795c05ee815a5da08 Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.293164 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-77dcf56988-s8mzf_721e1031-d80c-406a-ad29-17879d58c3c3/console/0.log" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.293328 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-77dcf56988-s8mzf" event={"ID":"721e1031-d80c-406a-ad29-17879d58c3c3","Type":"ContainerDied","Data":"f31650000a80adda0dbae2781d24e63e5ca5da99fa3ba17b6a2588f41cca02cd"} Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.293404 4824 scope.go:117] "RemoveContainer" containerID="88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.293682 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-77dcf56988-s8mzf" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.314883 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9177fe98-0bd2-43b4-85bb-f9218d48211e-dns-svc\") pod \"9177fe98-0bd2-43b4-85bb-f9218d48211e\" (UID: \"9177fe98-0bd2-43b4-85bb-f9218d48211e\") " Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.314986 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78pwt\" (UniqueName: \"kubernetes.io/projected/9177fe98-0bd2-43b4-85bb-f9218d48211e-kube-api-access-78pwt\") pod \"9177fe98-0bd2-43b4-85bb-f9218d48211e\" (UID: \"9177fe98-0bd2-43b4-85bb-f9218d48211e\") " Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.315103 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9177fe98-0bd2-43b4-85bb-f9218d48211e-config\") pod \"9177fe98-0bd2-43b4-85bb-f9218d48211e\" (UID: \"9177fe98-0bd2-43b4-85bb-f9218d48211e\") " Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.315162 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9177fe98-0bd2-43b4-85bb-f9218d48211e-ovsdbserver-sb\") pod \"9177fe98-0bd2-43b4-85bb-f9218d48211e\" (UID: \"9177fe98-0bd2-43b4-85bb-f9218d48211e\") " Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.328466 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-985bt-config-x2qzv" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.328652 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"07c1822c-1257-4d8c-9a17-d921929000fa","Type":"ContainerStarted","Data":"d35277030f4a65e4facfc048ac021fc3c08167f2f2aceae17375fa39c64384b6"} Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.329698 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.331065 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9177fe98-0bd2-43b4-85bb-f9218d48211e-kube-api-access-78pwt" (OuterVolumeSpecName: "kube-api-access-78pwt") pod "9177fe98-0bd2-43b4-85bb-f9218d48211e" (UID: "9177fe98-0bd2-43b4-85bb-f9218d48211e"). InnerVolumeSpecName "kube-api-access-78pwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.382732 4824 generic.go:334] "Generic (PLEG): container finished" podID="9177fe98-0bd2-43b4-85bb-f9218d48211e" containerID="d2d1b5606cbde06cfba9a3a183389179cd58d0f9d03ccc9ee2a7f41baaa97c94" exitCode=0 Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.384023 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" event={"ID":"9177fe98-0bd2-43b4-85bb-f9218d48211e","Type":"ContainerDied","Data":"d2d1b5606cbde06cfba9a3a183389179cd58d0f9d03ccc9ee2a7f41baaa97c94"} Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.384057 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" event={"ID":"9177fe98-0bd2-43b4-85bb-f9218d48211e","Type":"ContainerDied","Data":"4b087d5395fa6b4a6d06e08ddb1513837e0502acc753cd2a79cd585626b29941"} Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.383881 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-p4czz" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.396542 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-ccjw2" event={"ID":"245fbd50-9e68-416a-a260-56363b243f55","Type":"ContainerStarted","Data":"d962a2c4ff17cdd9b661c6b3df351c93b598c1a1f3a4a4c795c05ee815a5da08"} Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.407704 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b025-account-create-update-cxmdd" event={"ID":"373c7c35-5998-40d4-b40b-5024d81e5567","Type":"ContainerStarted","Data":"d37c75642aed323a19fc569cba2cfadc1c31c83a932ed4466444c6240731dc17"} Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.407734 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-ncg6n" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.408549 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=41.673391057 podStartE2EDuration="1m21.408525442s" podCreationTimestamp="2025-12-09 10:08:15 +0000 UTC" firstStartedPulling="2025-12-09 10:08:18.164637736 +0000 UTC m=+1254.499142403" lastFinishedPulling="2025-12-09 10:08:57.899772121 +0000 UTC m=+1294.234276788" observedRunningTime="2025-12-09 10:09:36.354907893 +0000 UTC m=+1332.689412560" watchObservedRunningTime="2025-12-09 10:09:36.408525442 +0000 UTC m=+1332.743030109" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.423350 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78pwt\" (UniqueName: \"kubernetes.io/projected/9177fe98-0bd2-43b4-85bb-f9218d48211e-kube-api-access-78pwt\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.423414 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9177fe98-0bd2-43b4-85bb-f9218d48211e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9177fe98-0bd2-43b4-85bb-f9218d48211e" (UID: "9177fe98-0bd2-43b4-85bb-f9218d48211e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.438383 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-77dcf56988-s8mzf"] Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.439936 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9177fe98-0bd2-43b4-85bb-f9218d48211e-config" (OuterVolumeSpecName: "config") pod "9177fe98-0bd2-43b4-85bb-f9218d48211e" (UID: "9177fe98-0bd2-43b4-85bb-f9218d48211e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.453404 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-77dcf56988-s8mzf"] Dec 09 10:09:36 crc kubenswrapper[4824]: W1209 10:09:36.466370 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ec0f64d_691b_4433_a92b_ae6ab0ae4971.slice/crio-e3814a4f6d3b506baf8d5ddfde1c7424e7afa74f1bab2f5b38f16974c82a14ed WatchSource:0}: Error finding container e3814a4f6d3b506baf8d5ddfde1c7424e7afa74f1bab2f5b38f16974c82a14ed: Status 404 returned error can't find the container with id e3814a4f6d3b506baf8d5ddfde1c7424e7afa74f1bab2f5b38f16974c82a14ed Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.466650 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-65e6-account-create-update-znnqb"] Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.481573 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9177fe98-0bd2-43b4-85bb-f9218d48211e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9177fe98-0bd2-43b4-85bb-f9218d48211e" (UID: "9177fe98-0bd2-43b4-85bb-f9218d48211e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.515389 4824 scope.go:117] "RemoveContainer" containerID="d2d1b5606cbde06cfba9a3a183389179cd58d0f9d03ccc9ee2a7f41baaa97c94" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.526030 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9177fe98-0bd2-43b4-85bb-f9218d48211e-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.526063 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9177fe98-0bd2-43b4-85bb-f9218d48211e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.526076 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9177fe98-0bd2-43b4-85bb-f9218d48211e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.529172 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-5vfks"] Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.546273 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 09 10:09:36 crc kubenswrapper[4824]: W1209 10:09:36.559089 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02d954cc_c1d4_45c8_86a2_fe001a222616.slice/crio-51685a80eee0e11a9a0efbc6c43deb8f1a774b32a70f29effb7ca2a5e968965a WatchSource:0}: Error finding container 51685a80eee0e11a9a0efbc6c43deb8f1a774b32a70f29effb7ca2a5e968965a: Status 404 returned error can't find the container with id 51685a80eee0e11a9a0efbc6c43deb8f1a774b32a70f29effb7ca2a5e968965a Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.672107 4824 scope.go:117] "RemoveContainer" containerID="3516d3c603a4d43ca10f7c358d86e94b214fe1665395df66d0bbae876f8b963b" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.717896 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-fmfsh"] Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.785547 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-7029-account-create-update-dkjjt"] Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.810227 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-ncg6n"] Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.829478 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-ncg6n"] Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.834055 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-etc-swift\") pod \"swift-storage-0\" (UID: \"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f\") " pod="openstack/swift-storage-0" Dec 09 10:09:36 crc kubenswrapper[4824]: E1209 10:09:36.834343 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 09 10:09:36 crc kubenswrapper[4824]: E1209 10:09:36.834375 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 09 10:09:36 crc kubenswrapper[4824]: E1209 10:09:36.834451 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-etc-swift podName:66fd04f1-a0d9-4295-ac42-e5b20acc7c1f nodeName:}" failed. No retries permitted until 2025-12-09 10:09:38.834429603 +0000 UTC m=+1335.168934260 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-etc-swift") pod "swift-storage-0" (UID: "66fd04f1-a0d9-4295-ac42-e5b20acc7c1f") : configmap "swift-ring-files" not found Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.847217 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-p4czz"] Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.860218 4824 scope.go:117] "RemoveContainer" containerID="d2d1b5606cbde06cfba9a3a183389179cd58d0f9d03ccc9ee2a7f41baaa97c94" Dec 09 10:09:36 crc kubenswrapper[4824]: E1209 10:09:36.862201 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2d1b5606cbde06cfba9a3a183389179cd58d0f9d03ccc9ee2a7f41baaa97c94\": container with ID starting with d2d1b5606cbde06cfba9a3a183389179cd58d0f9d03ccc9ee2a7f41baaa97c94 not found: ID does not exist" containerID="d2d1b5606cbde06cfba9a3a183389179cd58d0f9d03ccc9ee2a7f41baaa97c94" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.862252 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2d1b5606cbde06cfba9a3a183389179cd58d0f9d03ccc9ee2a7f41baaa97c94"} err="failed to get container status \"d2d1b5606cbde06cfba9a3a183389179cd58d0f9d03ccc9ee2a7f41baaa97c94\": rpc error: code = NotFound desc = could not find container \"d2d1b5606cbde06cfba9a3a183389179cd58d0f9d03ccc9ee2a7f41baaa97c94\": container with ID starting with d2d1b5606cbde06cfba9a3a183389179cd58d0f9d03ccc9ee2a7f41baaa97c94 not found: ID does not exist" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.862286 4824 scope.go:117] "RemoveContainer" containerID="3516d3c603a4d43ca10f7c358d86e94b214fe1665395df66d0bbae876f8b963b" Dec 09 10:09:36 crc kubenswrapper[4824]: E1209 10:09:36.868134 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3516d3c603a4d43ca10f7c358d86e94b214fe1665395df66d0bbae876f8b963b\": container with ID starting with 3516d3c603a4d43ca10f7c358d86e94b214fe1665395df66d0bbae876f8b963b not found: ID does not exist" containerID="3516d3c603a4d43ca10f7c358d86e94b214fe1665395df66d0bbae876f8b963b" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.868191 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3516d3c603a4d43ca10f7c358d86e94b214fe1665395df66d0bbae876f8b963b"} err="failed to get container status \"3516d3c603a4d43ca10f7c358d86e94b214fe1665395df66d0bbae876f8b963b\": rpc error: code = NotFound desc = could not find container \"3516d3c603a4d43ca10f7c358d86e94b214fe1665395df66d0bbae876f8b963b\": container with ID starting with 3516d3c603a4d43ca10f7c358d86e94b214fe1665395df66d0bbae876f8b963b not found: ID does not exist" Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.933367 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-p4czz"] Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.985724 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-a94f-account-create-update-f2dbj"] Dec 09 10:09:36 crc kubenswrapper[4824]: I1209 10:09:36.994251 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-t4gz7"] Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.097639 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-cf5wn"] Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.154816 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-pztmz"] Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.336192 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-985bt-config-x2qzv"] Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.424300 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-cf5wn" event={"ID":"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c","Type":"ContainerStarted","Data":"94d49d8097ae1fd76fdd4e47953421dfcf4ba9bd66f8dc14f0ffd65d153193df"} Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.426617 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b025-account-create-update-cxmdd" event={"ID":"373c7c35-5998-40d4-b40b-5024d81e5567","Type":"ContainerStarted","Data":"f5f9c6e449b6d579f129798f5c0a7d4f334e6eba1d5832bbd2c1e34b78142745"} Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.432452 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65e6-account-create-update-znnqb" event={"ID":"1ec0f64d-691b-4433-a92b-ae6ab0ae4971","Type":"ContainerStarted","Data":"6bb064295473fab572428ffb93bdfc3abbae6a6ba39d21f98cdb6d98d7a4700c"} Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.432498 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65e6-account-create-update-znnqb" event={"ID":"1ec0f64d-691b-4433-a92b-ae6ab0ae4971","Type":"ContainerStarted","Data":"e3814a4f6d3b506baf8d5ddfde1c7424e7afa74f1bab2f5b38f16974c82a14ed"} Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.436763 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a94f-account-create-update-f2dbj" event={"ID":"1c6345d1-110d-4671-b27b-f8d2abcf4d4f","Type":"ContainerStarted","Data":"49b988ff58f5cd3d4a3d26e92bd08ad81cfc30b85bab89c588f85088a19226b0"} Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.442166 4824 generic.go:334] "Generic (PLEG): container finished" podID="245fbd50-9e68-416a-a260-56363b243f55" containerID="d56cf72eb935f510b885c872fa34865ff360c27b73d0ff6b22950b29e79b50e1" exitCode=0 Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.442253 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-ccjw2" event={"ID":"245fbd50-9e68-416a-a260-56363b243f55","Type":"ContainerDied","Data":"d56cf72eb935f510b885c872fa34865ff360c27b73d0ff6b22950b29e79b50e1"} Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.452884 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-b025-account-create-update-cxmdd" podStartSLOduration=18.452867336 podStartE2EDuration="18.452867336s" podCreationTimestamp="2025-12-09 10:09:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:09:37.449023904 +0000 UTC m=+1333.783528581" watchObservedRunningTime="2025-12-09 10:09:37.452867336 +0000 UTC m=+1333.787371993" Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.464225 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-7029-account-create-update-dkjjt" event={"ID":"83cb8977-7594-4739-90ae-ba915822102e","Type":"ContainerStarted","Data":"ea6e20c5563c87a5c35987ea4d40a6bae863eea01a42d6e87ef1edb2fc2311b0"} Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.464299 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-7029-account-create-update-dkjjt" event={"ID":"83cb8977-7594-4739-90ae-ba915822102e","Type":"ContainerStarted","Data":"677cce326770685f100e61ea719b4a5c10e30f2bb026e6d3390cdd9d8a264aea"} Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.471591 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-65e6-account-create-update-znnqb" podStartSLOduration=18.471566149 podStartE2EDuration="18.471566149s" podCreationTimestamp="2025-12-09 10:09:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:09:37.468663236 +0000 UTC m=+1333.803167903" watchObservedRunningTime="2025-12-09 10:09:37.471566149 +0000 UTC m=+1333.806070816" Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.499922 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-t4gz7" event={"ID":"82ce77fc-cdbd-4df3-b067-195040a2836e","Type":"ContainerStarted","Data":"11ad6ca9d90a564448cdb01df3ff58d259af1eb8991df735eea3faabbe2714da"} Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.510218 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-985bt-config-x2qzv" event={"ID":"79e597f5-64c0-4250-b3d4-69ac3db9cb69","Type":"ContainerStarted","Data":"814d0e0ed563eed0c969bcbd23b0b17bdaf888ef28f75667a5220dab78d09037"} Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.513546 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-fmfsh" event={"ID":"033734aa-97c9-498f-8bae-992507b79f71","Type":"ContainerStarted","Data":"50499ef0bdf5865315ab6b6f9e726f19a093a2cf234f18fb4b15b596a33a6ee9"} Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.513603 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-fmfsh" event={"ID":"033734aa-97c9-498f-8bae-992507b79f71","Type":"ContainerStarted","Data":"af7c5ec4bda9c19df9cfaa4dea30203887db49dee36fc652d0f6798631367647"} Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.523413 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pztmz" event={"ID":"7b6a6023-0aa7-4778-be37-6de2222088c4","Type":"ContainerStarted","Data":"458f646b2f33e9818a1aa334ebb0d26f8533ac59ca8ee2b7575ae4c9a14f951e"} Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.540773 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-5vfks" event={"ID":"02d954cc-c1d4-45c8-86a2-fe001a222616","Type":"ContainerStarted","Data":"29daa08ae30f5804d16cd244260b909954b1a9f61e7b3822eaba0d7cc8642676"} Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.540829 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-5vfks" event={"ID":"02d954cc-c1d4-45c8-86a2-fe001a222616","Type":"ContainerStarted","Data":"51685a80eee0e11a9a0efbc6c43deb8f1a774b32a70f29effb7ca2a5e968965a"} Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.572508 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-7029-account-create-update-dkjjt" podStartSLOduration=16.572481997 podStartE2EDuration="16.572481997s" podCreationTimestamp="2025-12-09 10:09:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:09:37.523616649 +0000 UTC m=+1333.858121316" watchObservedRunningTime="2025-12-09 10:09:37.572481997 +0000 UTC m=+1333.906986654" Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.605228 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-openstack-db-create-fmfsh" podStartSLOduration=16.605204714 podStartE2EDuration="16.605204714s" podCreationTimestamp="2025-12-09 10:09:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:09:37.545282135 +0000 UTC m=+1333.879786802" watchObservedRunningTime="2025-12-09 10:09:37.605204714 +0000 UTC m=+1333.939709381" Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.933757 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="721e1031-d80c-406a-ad29-17879d58c3c3" path="/var/lib/kubelet/pods/721e1031-d80c-406a-ad29-17879d58c3c3/volumes" Dec 09 10:09:37 crc kubenswrapper[4824]: I1209 10:09:37.934631 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9177fe98-0bd2-43b4-85bb-f9218d48211e" path="/var/lib/kubelet/pods/9177fe98-0bd2-43b4-85bb-f9218d48211e/volumes" Dec 09 10:09:38 crc kubenswrapper[4824]: E1209 10:09:38.269641 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721e1031_d80c_406a_ad29_17879d58c3c3.slice/crio-88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c1822c_1257_4d8c_9a17_d921929000fa.slice/crio-conmon-b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721e1031_d80c_406a_ad29_17879d58c3c3.slice/crio-conmon-88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c1822c_1257_4d8c_9a17_d921929000fa.slice/crio-b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c.scope\": RecentStats: unable to find data in memory cache]" Dec 09 10:09:38 crc kubenswrapper[4824]: I1209 10:09:38.552604 4824 generic.go:334] "Generic (PLEG): container finished" podID="7b6a6023-0aa7-4778-be37-6de2222088c4" containerID="afa3e9b8e9f87b390281f45b6490b3d377461954b20ca3c6deaae6c025b61aa2" exitCode=0 Dec 09 10:09:38 crc kubenswrapper[4824]: I1209 10:09:38.552677 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pztmz" event={"ID":"7b6a6023-0aa7-4778-be37-6de2222088c4","Type":"ContainerDied","Data":"afa3e9b8e9f87b390281f45b6490b3d377461954b20ca3c6deaae6c025b61aa2"} Dec 09 10:09:38 crc kubenswrapper[4824]: I1209 10:09:38.560379 4824 generic.go:334] "Generic (PLEG): container finished" podID="83cb8977-7594-4739-90ae-ba915822102e" containerID="ea6e20c5563c87a5c35987ea4d40a6bae863eea01a42d6e87ef1edb2fc2311b0" exitCode=0 Dec 09 10:09:38 crc kubenswrapper[4824]: I1209 10:09:38.560449 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-7029-account-create-update-dkjjt" event={"ID":"83cb8977-7594-4739-90ae-ba915822102e","Type":"ContainerDied","Data":"ea6e20c5563c87a5c35987ea4d40a6bae863eea01a42d6e87ef1edb2fc2311b0"} Dec 09 10:09:38 crc kubenswrapper[4824]: I1209 10:09:38.564688 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-985bt-config-x2qzv" event={"ID":"79e597f5-64c0-4250-b3d4-69ac3db9cb69","Type":"ContainerStarted","Data":"0a579075bdc0f1dc087391998446e9f4e87f86205adca94cd10a32f6dcd5b13b"} Dec 09 10:09:38 crc kubenswrapper[4824]: I1209 10:09:38.574368 4824 generic.go:334] "Generic (PLEG): container finished" podID="82ce77fc-cdbd-4df3-b067-195040a2836e" containerID="0123721178eca4cbfc6b2f6749af4123c08e9b695c4bd8681e0c0fc9b3126303" exitCode=0 Dec 09 10:09:38 crc kubenswrapper[4824]: I1209 10:09:38.574470 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-t4gz7" event={"ID":"82ce77fc-cdbd-4df3-b067-195040a2836e","Type":"ContainerDied","Data":"0123721178eca4cbfc6b2f6749af4123c08e9b695c4bd8681e0c0fc9b3126303"} Dec 09 10:09:38 crc kubenswrapper[4824]: I1209 10:09:38.576076 4824 generic.go:334] "Generic (PLEG): container finished" podID="033734aa-97c9-498f-8bae-992507b79f71" containerID="50499ef0bdf5865315ab6b6f9e726f19a093a2cf234f18fb4b15b596a33a6ee9" exitCode=0 Dec 09 10:09:38 crc kubenswrapper[4824]: I1209 10:09:38.576162 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-fmfsh" event={"ID":"033734aa-97c9-498f-8bae-992507b79f71","Type":"ContainerDied","Data":"50499ef0bdf5865315ab6b6f9e726f19a093a2cf234f18fb4b15b596a33a6ee9"} Dec 09 10:09:38 crc kubenswrapper[4824]: I1209 10:09:38.579517 4824 generic.go:334] "Generic (PLEG): container finished" podID="1ec0f64d-691b-4433-a92b-ae6ab0ae4971" containerID="6bb064295473fab572428ffb93bdfc3abbae6a6ba39d21f98cdb6d98d7a4700c" exitCode=0 Dec 09 10:09:38 crc kubenswrapper[4824]: I1209 10:09:38.579623 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65e6-account-create-update-znnqb" event={"ID":"1ec0f64d-691b-4433-a92b-ae6ab0ae4971","Type":"ContainerDied","Data":"6bb064295473fab572428ffb93bdfc3abbae6a6ba39d21f98cdb6d98d7a4700c"} Dec 09 10:09:38 crc kubenswrapper[4824]: I1209 10:09:38.582676 4824 generic.go:334] "Generic (PLEG): container finished" podID="02d954cc-c1d4-45c8-86a2-fe001a222616" containerID="29daa08ae30f5804d16cd244260b909954b1a9f61e7b3822eaba0d7cc8642676" exitCode=0 Dec 09 10:09:38 crc kubenswrapper[4824]: I1209 10:09:38.582757 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-5vfks" event={"ID":"02d954cc-c1d4-45c8-86a2-fe001a222616","Type":"ContainerDied","Data":"29daa08ae30f5804d16cd244260b909954b1a9f61e7b3822eaba0d7cc8642676"} Dec 09 10:09:38 crc kubenswrapper[4824]: I1209 10:09:38.586423 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a94f-account-create-update-f2dbj" event={"ID":"1c6345d1-110d-4671-b27b-f8d2abcf4d4f","Type":"ContainerStarted","Data":"17314388a85fcca06ec1542f5802fb055e6225086fc3e4296338e07364427259"} Dec 09 10:09:38 crc kubenswrapper[4824]: I1209 10:09:38.598790 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-985bt-config-x2qzv" podStartSLOduration=3.5987565679999998 podStartE2EDuration="3.598756568s" podCreationTimestamp="2025-12-09 10:09:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:09:38.586562382 +0000 UTC m=+1334.921067049" watchObservedRunningTime="2025-12-09 10:09:38.598756568 +0000 UTC m=+1334.933261235" Dec 09 10:09:38 crc kubenswrapper[4824]: I1209 10:09:38.601128 4824 generic.go:334] "Generic (PLEG): container finished" podID="373c7c35-5998-40d4-b40b-5024d81e5567" containerID="f5f9c6e449b6d579f129798f5c0a7d4f334e6eba1d5832bbd2c1e34b78142745" exitCode=0 Dec 09 10:09:38 crc kubenswrapper[4824]: I1209 10:09:38.601236 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b025-account-create-update-cxmdd" event={"ID":"373c7c35-5998-40d4-b40b-5024d81e5567","Type":"ContainerDied","Data":"f5f9c6e449b6d579f129798f5c0a7d4f334e6eba1d5832bbd2c1e34b78142745"} Dec 09 10:09:38 crc kubenswrapper[4824]: I1209 10:09:38.852023 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-etc-swift\") pod \"swift-storage-0\" (UID: \"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f\") " pod="openstack/swift-storage-0" Dec 09 10:09:38 crc kubenswrapper[4824]: E1209 10:09:38.852215 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 09 10:09:38 crc kubenswrapper[4824]: E1209 10:09:38.852698 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 09 10:09:38 crc kubenswrapper[4824]: E1209 10:09:38.852763 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-etc-swift podName:66fd04f1-a0d9-4295-ac42-e5b20acc7c1f nodeName:}" failed. No retries permitted until 2025-12-09 10:09:42.85273965 +0000 UTC m=+1339.187244317 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-etc-swift") pod "swift-storage-0" (UID: "66fd04f1-a0d9-4295-ac42-e5b20acc7c1f") : configmap "swift-ring-files" not found Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.223295 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-5vfks" Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.234604 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ccjw2" Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.263622 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/245fbd50-9e68-416a-a260-56363b243f55-operator-scripts\") pod \"245fbd50-9e68-416a-a260-56363b243f55\" (UID: \"245fbd50-9e68-416a-a260-56363b243f55\") " Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.263935 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdl7p\" (UniqueName: \"kubernetes.io/projected/02d954cc-c1d4-45c8-86a2-fe001a222616-kube-api-access-mdl7p\") pod \"02d954cc-c1d4-45c8-86a2-fe001a222616\" (UID: \"02d954cc-c1d4-45c8-86a2-fe001a222616\") " Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.264024 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/02d954cc-c1d4-45c8-86a2-fe001a222616-operator-scripts\") pod \"02d954cc-c1d4-45c8-86a2-fe001a222616\" (UID: \"02d954cc-c1d4-45c8-86a2-fe001a222616\") " Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.264067 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgm5w\" (UniqueName: \"kubernetes.io/projected/245fbd50-9e68-416a-a260-56363b243f55-kube-api-access-lgm5w\") pod \"245fbd50-9e68-416a-a260-56363b243f55\" (UID: \"245fbd50-9e68-416a-a260-56363b243f55\") " Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.264933 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/245fbd50-9e68-416a-a260-56363b243f55-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "245fbd50-9e68-416a-a260-56363b243f55" (UID: "245fbd50-9e68-416a-a260-56363b243f55"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.264961 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02d954cc-c1d4-45c8-86a2-fe001a222616-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "02d954cc-c1d4-45c8-86a2-fe001a222616" (UID: "02d954cc-c1d4-45c8-86a2-fe001a222616"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.270028 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02d954cc-c1d4-45c8-86a2-fe001a222616-kube-api-access-mdl7p" (OuterVolumeSpecName: "kube-api-access-mdl7p") pod "02d954cc-c1d4-45c8-86a2-fe001a222616" (UID: "02d954cc-c1d4-45c8-86a2-fe001a222616"). InnerVolumeSpecName "kube-api-access-mdl7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.271453 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/245fbd50-9e68-416a-a260-56363b243f55-kube-api-access-lgm5w" (OuterVolumeSpecName: "kube-api-access-lgm5w") pod "245fbd50-9e68-416a-a260-56363b243f55" (UID: "245fbd50-9e68-416a-a260-56363b243f55"). InnerVolumeSpecName "kube-api-access-lgm5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.370761 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/245fbd50-9e68-416a-a260-56363b243f55-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.370816 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdl7p\" (UniqueName: \"kubernetes.io/projected/02d954cc-c1d4-45c8-86a2-fe001a222616-kube-api-access-mdl7p\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.370831 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/02d954cc-c1d4-45c8-86a2-fe001a222616-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.370991 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgm5w\" (UniqueName: \"kubernetes.io/projected/245fbd50-9e68-416a-a260-56363b243f55-kube-api-access-lgm5w\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.617341 4824 generic.go:334] "Generic (PLEG): container finished" podID="79e597f5-64c0-4250-b3d4-69ac3db9cb69" containerID="0a579075bdc0f1dc087391998446e9f4e87f86205adca94cd10a32f6dcd5b13b" exitCode=0 Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.617428 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-985bt-config-x2qzv" event={"ID":"79e597f5-64c0-4250-b3d4-69ac3db9cb69","Type":"ContainerDied","Data":"0a579075bdc0f1dc087391998446e9f4e87f86205adca94cd10a32f6dcd5b13b"} Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.619828 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-5vfks" event={"ID":"02d954cc-c1d4-45c8-86a2-fe001a222616","Type":"ContainerDied","Data":"51685a80eee0e11a9a0efbc6c43deb8f1a774b32a70f29effb7ca2a5e968965a"} Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.619886 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51685a80eee0e11a9a0efbc6c43deb8f1a774b32a70f29effb7ca2a5e968965a" Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.619842 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-5vfks" Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.625692 4824 generic.go:334] "Generic (PLEG): container finished" podID="1c6345d1-110d-4671-b27b-f8d2abcf4d4f" containerID="17314388a85fcca06ec1542f5802fb055e6225086fc3e4296338e07364427259" exitCode=0 Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.625790 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a94f-account-create-update-f2dbj" event={"ID":"1c6345d1-110d-4671-b27b-f8d2abcf4d4f","Type":"ContainerDied","Data":"17314388a85fcca06ec1542f5802fb055e6225086fc3e4296338e07364427259"} Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.635195 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ccjw2" Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.635429 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-ccjw2" event={"ID":"245fbd50-9e68-416a-a260-56363b243f55","Type":"ContainerDied","Data":"d962a2c4ff17cdd9b661c6b3df351c93b598c1a1f3a4a4c795c05ee815a5da08"} Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.635478 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d962a2c4ff17cdd9b661c6b3df351c93b598c1a1f3a4a4c795c05ee815a5da08" Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.651830 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-t4gz7" event={"ID":"82ce77fc-cdbd-4df3-b067-195040a2836e","Type":"ContainerStarted","Data":"2b89f32a2604297e7699c708001e61817fa58b671867a19ac510149bc6732369"} Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.652611 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-t4gz7" Dec 09 10:09:39 crc kubenswrapper[4824]: I1209 10:09:39.701308 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-t4gz7" podStartSLOduration=18.701278567 podStartE2EDuration="18.701278567s" podCreationTimestamp="2025-12-09 10:09:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:09:39.692879921 +0000 UTC m=+1336.027384608" watchObservedRunningTime="2025-12-09 10:09:39.701278567 +0000 UTC m=+1336.035783234" Dec 09 10:09:40 crc kubenswrapper[4824]: I1209 10:09:40.553231 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-985bt" Dec 09 10:09:42 crc kubenswrapper[4824]: I1209 10:09:42.875270 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-etc-swift\") pod \"swift-storage-0\" (UID: \"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f\") " pod="openstack/swift-storage-0" Dec 09 10:09:42 crc kubenswrapper[4824]: E1209 10:09:42.875797 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 09 10:09:42 crc kubenswrapper[4824]: E1209 10:09:42.877336 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 09 10:09:42 crc kubenswrapper[4824]: E1209 10:09:42.877414 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-etc-swift podName:66fd04f1-a0d9-4295-ac42-e5b20acc7c1f nodeName:}" failed. No retries permitted until 2025-12-09 10:09:50.877383934 +0000 UTC m=+1347.211888611 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-etc-swift") pod "swift-storage-0" (UID: "66fd04f1-a0d9-4295-ac42-e5b20acc7c1f") : configmap "swift-ring-files" not found Dec 09 10:09:42 crc kubenswrapper[4824]: I1209 10:09:42.882319 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-7029-account-create-update-dkjjt" Dec 09 10:09:42 crc kubenswrapper[4824]: I1209 10:09:42.890945 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pztmz" Dec 09 10:09:42 crc kubenswrapper[4824]: I1209 10:09:42.921357 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a94f-account-create-update-f2dbj" Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.242168 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxp2s\" (UniqueName: \"kubernetes.io/projected/7b6a6023-0aa7-4778-be37-6de2222088c4-kube-api-access-xxp2s\") pod \"7b6a6023-0aa7-4778-be37-6de2222088c4\" (UID: \"7b6a6023-0aa7-4778-be37-6de2222088c4\") " Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.242252 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c6345d1-110d-4671-b27b-f8d2abcf4d4f-operator-scripts\") pod \"1c6345d1-110d-4671-b27b-f8d2abcf4d4f\" (UID: \"1c6345d1-110d-4671-b27b-f8d2abcf4d4f\") " Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.242312 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83cb8977-7594-4739-90ae-ba915822102e-operator-scripts\") pod \"83cb8977-7594-4739-90ae-ba915822102e\" (UID: \"83cb8977-7594-4739-90ae-ba915822102e\") " Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.242370 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h85hd\" (UniqueName: \"kubernetes.io/projected/1c6345d1-110d-4671-b27b-f8d2abcf4d4f-kube-api-access-h85hd\") pod \"1c6345d1-110d-4671-b27b-f8d2abcf4d4f\" (UID: \"1c6345d1-110d-4671-b27b-f8d2abcf4d4f\") " Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.242572 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfcbm\" (UniqueName: \"kubernetes.io/projected/83cb8977-7594-4739-90ae-ba915822102e-kube-api-access-cfcbm\") pod \"83cb8977-7594-4739-90ae-ba915822102e\" (UID: \"83cb8977-7594-4739-90ae-ba915822102e\") " Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.242642 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b6a6023-0aa7-4778-be37-6de2222088c4-operator-scripts\") pod \"7b6a6023-0aa7-4778-be37-6de2222088c4\" (UID: \"7b6a6023-0aa7-4778-be37-6de2222088c4\") " Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.243692 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83cb8977-7594-4739-90ae-ba915822102e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "83cb8977-7594-4739-90ae-ba915822102e" (UID: "83cb8977-7594-4739-90ae-ba915822102e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.244683 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c6345d1-110d-4671-b27b-f8d2abcf4d4f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1c6345d1-110d-4671-b27b-f8d2abcf4d4f" (UID: "1c6345d1-110d-4671-b27b-f8d2abcf4d4f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.245760 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b6a6023-0aa7-4778-be37-6de2222088c4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7b6a6023-0aa7-4778-be37-6de2222088c4" (UID: "7b6a6023-0aa7-4778-be37-6de2222088c4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.251344 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83cb8977-7594-4739-90ae-ba915822102e-kube-api-access-cfcbm" (OuterVolumeSpecName: "kube-api-access-cfcbm") pod "83cb8977-7594-4739-90ae-ba915822102e" (UID: "83cb8977-7594-4739-90ae-ba915822102e"). InnerVolumeSpecName "kube-api-access-cfcbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.252553 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c6345d1-110d-4671-b27b-f8d2abcf4d4f-kube-api-access-h85hd" (OuterVolumeSpecName: "kube-api-access-h85hd") pod "1c6345d1-110d-4671-b27b-f8d2abcf4d4f" (UID: "1c6345d1-110d-4671-b27b-f8d2abcf4d4f"). InnerVolumeSpecName "kube-api-access-h85hd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.253936 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b6a6023-0aa7-4778-be37-6de2222088c4-kube-api-access-xxp2s" (OuterVolumeSpecName: "kube-api-access-xxp2s") pod "7b6a6023-0aa7-4778-be37-6de2222088c4" (UID: "7b6a6023-0aa7-4778-be37-6de2222088c4"). InnerVolumeSpecName "kube-api-access-xxp2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.346393 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfcbm\" (UniqueName: \"kubernetes.io/projected/83cb8977-7594-4739-90ae-ba915822102e-kube-api-access-cfcbm\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.346751 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b6a6023-0aa7-4778-be37-6de2222088c4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.346764 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxp2s\" (UniqueName: \"kubernetes.io/projected/7b6a6023-0aa7-4778-be37-6de2222088c4-kube-api-access-xxp2s\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.346775 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c6345d1-110d-4671-b27b-f8d2abcf4d4f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.346876 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83cb8977-7594-4739-90ae-ba915822102e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.346888 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h85hd\" (UniqueName: \"kubernetes.io/projected/1c6345d1-110d-4671-b27b-f8d2abcf4d4f-kube-api-access-h85hd\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.719567 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pztmz" Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.719602 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pztmz" event={"ID":"7b6a6023-0aa7-4778-be37-6de2222088c4","Type":"ContainerDied","Data":"458f646b2f33e9818a1aa334ebb0d26f8533ac59ca8ee2b7575ae4c9a14f951e"} Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.719652 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="458f646b2f33e9818a1aa334ebb0d26f8533ac59ca8ee2b7575ae4c9a14f951e" Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.723443 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a94f-account-create-update-f2dbj" event={"ID":"1c6345d1-110d-4671-b27b-f8d2abcf4d4f","Type":"ContainerDied","Data":"49b988ff58f5cd3d4a3d26e92bd08ad81cfc30b85bab89c588f85088a19226b0"} Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.723484 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49b988ff58f5cd3d4a3d26e92bd08ad81cfc30b85bab89c588f85088a19226b0" Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.723569 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a94f-account-create-update-f2dbj" Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.728136 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-7029-account-create-update-dkjjt" event={"ID":"83cb8977-7594-4739-90ae-ba915822102e","Type":"ContainerDied","Data":"677cce326770685f100e61ea719b4a5c10e30f2bb026e6d3390cdd9d8a264aea"} Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.728165 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="677cce326770685f100e61ea719b4a5c10e30f2bb026e6d3390cdd9d8a264aea" Dec 09 10:09:43 crc kubenswrapper[4824]: I1209 10:09:43.728270 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-7029-account-create-update-dkjjt" Dec 09 10:09:44 crc kubenswrapper[4824]: I1209 10:09:44.929185 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-65e6-account-create-update-znnqb" Dec 09 10:09:44 crc kubenswrapper[4824]: I1209 10:09:44.936877 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-fmfsh" Dec 09 10:09:44 crc kubenswrapper[4824]: I1209 10:09:44.957877 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-985bt-config-x2qzv" Dec 09 10:09:44 crc kubenswrapper[4824]: I1209 10:09:44.962200 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b025-account-create-update-cxmdd" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.086972 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5p9f9\" (UniqueName: \"kubernetes.io/projected/1ec0f64d-691b-4433-a92b-ae6ab0ae4971-kube-api-access-5p9f9\") pod \"1ec0f64d-691b-4433-a92b-ae6ab0ae4971\" (UID: \"1ec0f64d-691b-4433-a92b-ae6ab0ae4971\") " Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.087022 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2d7v\" (UniqueName: \"kubernetes.io/projected/373c7c35-5998-40d4-b40b-5024d81e5567-kube-api-access-r2d7v\") pod \"373c7c35-5998-40d4-b40b-5024d81e5567\" (UID: \"373c7c35-5998-40d4-b40b-5024d81e5567\") " Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.087068 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79e597f5-64c0-4250-b3d4-69ac3db9cb69-scripts\") pod \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\" (UID: \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\") " Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.087132 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/033734aa-97c9-498f-8bae-992507b79f71-operator-scripts\") pod \"033734aa-97c9-498f-8bae-992507b79f71\" (UID: \"033734aa-97c9-498f-8bae-992507b79f71\") " Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.087162 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2vhw\" (UniqueName: \"kubernetes.io/projected/79e597f5-64c0-4250-b3d4-69ac3db9cb69-kube-api-access-f2vhw\") pod \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\" (UID: \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\") " Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.087248 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/79e597f5-64c0-4250-b3d4-69ac3db9cb69-additional-scripts\") pod \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\" (UID: \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\") " Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.087309 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ec0f64d-691b-4433-a92b-ae6ab0ae4971-operator-scripts\") pod \"1ec0f64d-691b-4433-a92b-ae6ab0ae4971\" (UID: \"1ec0f64d-691b-4433-a92b-ae6ab0ae4971\") " Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.087377 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/373c7c35-5998-40d4-b40b-5024d81e5567-operator-scripts\") pod \"373c7c35-5998-40d4-b40b-5024d81e5567\" (UID: \"373c7c35-5998-40d4-b40b-5024d81e5567\") " Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.087454 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/79e597f5-64c0-4250-b3d4-69ac3db9cb69-var-run-ovn\") pod \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\" (UID: \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\") " Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.087491 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/79e597f5-64c0-4250-b3d4-69ac3db9cb69-var-log-ovn\") pod \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\" (UID: \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\") " Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.087519 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/79e597f5-64c0-4250-b3d4-69ac3db9cb69-var-run\") pod \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\" (UID: \"79e597f5-64c0-4250-b3d4-69ac3db9cb69\") " Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.087547 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rq58n\" (UniqueName: \"kubernetes.io/projected/033734aa-97c9-498f-8bae-992507b79f71-kube-api-access-rq58n\") pod \"033734aa-97c9-498f-8bae-992507b79f71\" (UID: \"033734aa-97c9-498f-8bae-992507b79f71\") " Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.088928 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79e597f5-64c0-4250-b3d4-69ac3db9cb69-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "79e597f5-64c0-4250-b3d4-69ac3db9cb69" (UID: "79e597f5-64c0-4250-b3d4-69ac3db9cb69"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.089987 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/033734aa-97c9-498f-8bae-992507b79f71-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "033734aa-97c9-498f-8bae-992507b79f71" (UID: "033734aa-97c9-498f-8bae-992507b79f71"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.089991 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ec0f64d-691b-4433-a92b-ae6ab0ae4971-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1ec0f64d-691b-4433-a92b-ae6ab0ae4971" (UID: "1ec0f64d-691b-4433-a92b-ae6ab0ae4971"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.090018 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/79e597f5-64c0-4250-b3d4-69ac3db9cb69-var-run" (OuterVolumeSpecName: "var-run") pod "79e597f5-64c0-4250-b3d4-69ac3db9cb69" (UID: "79e597f5-64c0-4250-b3d4-69ac3db9cb69"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.090017 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/79e597f5-64c0-4250-b3d4-69ac3db9cb69-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "79e597f5-64c0-4250-b3d4-69ac3db9cb69" (UID: "79e597f5-64c0-4250-b3d4-69ac3db9cb69"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.090041 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/79e597f5-64c0-4250-b3d4-69ac3db9cb69-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "79e597f5-64c0-4250-b3d4-69ac3db9cb69" (UID: "79e597f5-64c0-4250-b3d4-69ac3db9cb69"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.090764 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/373c7c35-5998-40d4-b40b-5024d81e5567-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "373c7c35-5998-40d4-b40b-5024d81e5567" (UID: "373c7c35-5998-40d4-b40b-5024d81e5567"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.091606 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79e597f5-64c0-4250-b3d4-69ac3db9cb69-scripts" (OuterVolumeSpecName: "scripts") pod "79e597f5-64c0-4250-b3d4-69ac3db9cb69" (UID: "79e597f5-64c0-4250-b3d4-69ac3db9cb69"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.101487 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ec0f64d-691b-4433-a92b-ae6ab0ae4971-kube-api-access-5p9f9" (OuterVolumeSpecName: "kube-api-access-5p9f9") pod "1ec0f64d-691b-4433-a92b-ae6ab0ae4971" (UID: "1ec0f64d-691b-4433-a92b-ae6ab0ae4971"). InnerVolumeSpecName "kube-api-access-5p9f9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.101680 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/373c7c35-5998-40d4-b40b-5024d81e5567-kube-api-access-r2d7v" (OuterVolumeSpecName: "kube-api-access-r2d7v") pod "373c7c35-5998-40d4-b40b-5024d81e5567" (UID: "373c7c35-5998-40d4-b40b-5024d81e5567"). InnerVolumeSpecName "kube-api-access-r2d7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.101772 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79e597f5-64c0-4250-b3d4-69ac3db9cb69-kube-api-access-f2vhw" (OuterVolumeSpecName: "kube-api-access-f2vhw") pod "79e597f5-64c0-4250-b3d4-69ac3db9cb69" (UID: "79e597f5-64c0-4250-b3d4-69ac3db9cb69"). InnerVolumeSpecName "kube-api-access-f2vhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.101906 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/033734aa-97c9-498f-8bae-992507b79f71-kube-api-access-rq58n" (OuterVolumeSpecName: "kube-api-access-rq58n") pod "033734aa-97c9-498f-8bae-992507b79f71" (UID: "033734aa-97c9-498f-8bae-992507b79f71"). InnerVolumeSpecName "kube-api-access-rq58n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.286586 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5p9f9\" (UniqueName: \"kubernetes.io/projected/1ec0f64d-691b-4433-a92b-ae6ab0ae4971-kube-api-access-5p9f9\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.286631 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2d7v\" (UniqueName: \"kubernetes.io/projected/373c7c35-5998-40d4-b40b-5024d81e5567-kube-api-access-r2d7v\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.286641 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79e597f5-64c0-4250-b3d4-69ac3db9cb69-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.286651 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/033734aa-97c9-498f-8bae-992507b79f71-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.286661 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2vhw\" (UniqueName: \"kubernetes.io/projected/79e597f5-64c0-4250-b3d4-69ac3db9cb69-kube-api-access-f2vhw\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.286669 4824 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/79e597f5-64c0-4250-b3d4-69ac3db9cb69-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.286677 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ec0f64d-691b-4433-a92b-ae6ab0ae4971-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.286714 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/373c7c35-5998-40d4-b40b-5024d81e5567-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.286722 4824 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/79e597f5-64c0-4250-b3d4-69ac3db9cb69-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.286730 4824 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/79e597f5-64c0-4250-b3d4-69ac3db9cb69-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.286738 4824 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/79e597f5-64c0-4250-b3d4-69ac3db9cb69-var-run\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.286746 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rq58n\" (UniqueName: \"kubernetes.io/projected/033734aa-97c9-498f-8bae-992507b79f71-kube-api-access-rq58n\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.494642 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-wbxqj"] Dec 09 10:09:45 crc kubenswrapper[4824]: E1209 10:09:45.496104 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02d954cc-c1d4-45c8-86a2-fe001a222616" containerName="mariadb-database-create" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.496129 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="02d954cc-c1d4-45c8-86a2-fe001a222616" containerName="mariadb-database-create" Dec 09 10:09:45 crc kubenswrapper[4824]: E1209 10:09:45.496141 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c6345d1-110d-4671-b27b-f8d2abcf4d4f" containerName="mariadb-account-create-update" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.496154 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c6345d1-110d-4671-b27b-f8d2abcf4d4f" containerName="mariadb-account-create-update" Dec 09 10:09:45 crc kubenswrapper[4824]: E1209 10:09:45.496173 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9177fe98-0bd2-43b4-85bb-f9218d48211e" containerName="dnsmasq-dns" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.496182 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9177fe98-0bd2-43b4-85bb-f9218d48211e" containerName="dnsmasq-dns" Dec 09 10:09:45 crc kubenswrapper[4824]: E1209 10:09:45.497479 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b6a6023-0aa7-4778-be37-6de2222088c4" containerName="mariadb-database-create" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.497492 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b6a6023-0aa7-4778-be37-6de2222088c4" containerName="mariadb-database-create" Dec 09 10:09:45 crc kubenswrapper[4824]: E1209 10:09:45.497523 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="373c7c35-5998-40d4-b40b-5024d81e5567" containerName="mariadb-account-create-update" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.497532 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="373c7c35-5998-40d4-b40b-5024d81e5567" containerName="mariadb-account-create-update" Dec 09 10:09:45 crc kubenswrapper[4824]: E1209 10:09:45.497547 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83cb8977-7594-4739-90ae-ba915822102e" containerName="mariadb-account-create-update" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.497554 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="83cb8977-7594-4739-90ae-ba915822102e" containerName="mariadb-account-create-update" Dec 09 10:09:45 crc kubenswrapper[4824]: E1209 10:09:45.497570 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="033734aa-97c9-498f-8bae-992507b79f71" containerName="mariadb-database-create" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.497577 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="033734aa-97c9-498f-8bae-992507b79f71" containerName="mariadb-database-create" Dec 09 10:09:45 crc kubenswrapper[4824]: E1209 10:09:45.497600 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="721e1031-d80c-406a-ad29-17879d58c3c3" containerName="console" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.497608 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="721e1031-d80c-406a-ad29-17879d58c3c3" containerName="console" Dec 09 10:09:45 crc kubenswrapper[4824]: E1209 10:09:45.497620 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ec0f64d-691b-4433-a92b-ae6ab0ae4971" containerName="mariadb-account-create-update" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.497627 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ec0f64d-691b-4433-a92b-ae6ab0ae4971" containerName="mariadb-account-create-update" Dec 09 10:09:45 crc kubenswrapper[4824]: E1209 10:09:45.497649 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79e597f5-64c0-4250-b3d4-69ac3db9cb69" containerName="ovn-config" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.497655 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="79e597f5-64c0-4250-b3d4-69ac3db9cb69" containerName="ovn-config" Dec 09 10:09:45 crc kubenswrapper[4824]: E1209 10:09:45.497672 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="245fbd50-9e68-416a-a260-56363b243f55" containerName="mariadb-database-create" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.497680 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="245fbd50-9e68-416a-a260-56363b243f55" containerName="mariadb-database-create" Dec 09 10:09:45 crc kubenswrapper[4824]: E1209 10:09:45.497697 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9177fe98-0bd2-43b4-85bb-f9218d48211e" containerName="init" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.497705 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9177fe98-0bd2-43b4-85bb-f9218d48211e" containerName="init" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.498027 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="02d954cc-c1d4-45c8-86a2-fe001a222616" containerName="mariadb-database-create" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.498049 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="373c7c35-5998-40d4-b40b-5024d81e5567" containerName="mariadb-account-create-update" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.498066 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="721e1031-d80c-406a-ad29-17879d58c3c3" containerName="console" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.498084 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="245fbd50-9e68-416a-a260-56363b243f55" containerName="mariadb-database-create" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.498098 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c6345d1-110d-4671-b27b-f8d2abcf4d4f" containerName="mariadb-account-create-update" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.498111 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ec0f64d-691b-4433-a92b-ae6ab0ae4971" containerName="mariadb-account-create-update" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.498134 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="79e597f5-64c0-4250-b3d4-69ac3db9cb69" containerName="ovn-config" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.498146 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b6a6023-0aa7-4778-be37-6de2222088c4" containerName="mariadb-database-create" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.498158 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="83cb8977-7594-4739-90ae-ba915822102e" containerName="mariadb-account-create-update" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.498172 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9177fe98-0bd2-43b4-85bb-f9218d48211e" containerName="dnsmasq-dns" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.498180 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="033734aa-97c9-498f-8bae-992507b79f71" containerName="mariadb-database-create" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.500076 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-wbxqj" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.503703 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-vpbz2" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.503893 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.514302 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-wbxqj"] Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.702971 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80a3aebf-6b04-4e56-81cc-a632b53293b6-config-data\") pod \"glance-db-sync-wbxqj\" (UID: \"80a3aebf-6b04-4e56-81cc-a632b53293b6\") " pod="openstack/glance-db-sync-wbxqj" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.703191 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6gxs\" (UniqueName: \"kubernetes.io/projected/80a3aebf-6b04-4e56-81cc-a632b53293b6-kube-api-access-p6gxs\") pod \"glance-db-sync-wbxqj\" (UID: \"80a3aebf-6b04-4e56-81cc-a632b53293b6\") " pod="openstack/glance-db-sync-wbxqj" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.703347 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/80a3aebf-6b04-4e56-81cc-a632b53293b6-db-sync-config-data\") pod \"glance-db-sync-wbxqj\" (UID: \"80a3aebf-6b04-4e56-81cc-a632b53293b6\") " pod="openstack/glance-db-sync-wbxqj" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.703592 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80a3aebf-6b04-4e56-81cc-a632b53293b6-combined-ca-bundle\") pod \"glance-db-sync-wbxqj\" (UID: \"80a3aebf-6b04-4e56-81cc-a632b53293b6\") " pod="openstack/glance-db-sync-wbxqj" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.760018 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-t4gz7" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.809140 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6gxs\" (UniqueName: \"kubernetes.io/projected/80a3aebf-6b04-4e56-81cc-a632b53293b6-kube-api-access-p6gxs\") pod \"glance-db-sync-wbxqj\" (UID: \"80a3aebf-6b04-4e56-81cc-a632b53293b6\") " pod="openstack/glance-db-sync-wbxqj" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.809227 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/80a3aebf-6b04-4e56-81cc-a632b53293b6-db-sync-config-data\") pod \"glance-db-sync-wbxqj\" (UID: \"80a3aebf-6b04-4e56-81cc-a632b53293b6\") " pod="openstack/glance-db-sync-wbxqj" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.809350 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80a3aebf-6b04-4e56-81cc-a632b53293b6-combined-ca-bundle\") pod \"glance-db-sync-wbxqj\" (UID: \"80a3aebf-6b04-4e56-81cc-a632b53293b6\") " pod="openstack/glance-db-sync-wbxqj" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.809420 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80a3aebf-6b04-4e56-81cc-a632b53293b6-config-data\") pod \"glance-db-sync-wbxqj\" (UID: \"80a3aebf-6b04-4e56-81cc-a632b53293b6\") " pod="openstack/glance-db-sync-wbxqj" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.831832 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80a3aebf-6b04-4e56-81cc-a632b53293b6-config-data\") pod \"glance-db-sync-wbxqj\" (UID: \"80a3aebf-6b04-4e56-81cc-a632b53293b6\") " pod="openstack/glance-db-sync-wbxqj" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.843825 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/80a3aebf-6b04-4e56-81cc-a632b53293b6-db-sync-config-data\") pod \"glance-db-sync-wbxqj\" (UID: \"80a3aebf-6b04-4e56-81cc-a632b53293b6\") " pod="openstack/glance-db-sync-wbxqj" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.848214 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-985bt-config-x2qzv" event={"ID":"79e597f5-64c0-4250-b3d4-69ac3db9cb69","Type":"ContainerDied","Data":"814d0e0ed563eed0c969bcbd23b0b17bdaf888ef28f75667a5220dab78d09037"} Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.848258 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="814d0e0ed563eed0c969bcbd23b0b17bdaf888ef28f75667a5220dab78d09037" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.848337 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-985bt-config-x2qzv" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.853179 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80a3aebf-6b04-4e56-81cc-a632b53293b6-combined-ca-bundle\") pod \"glance-db-sync-wbxqj\" (UID: \"80a3aebf-6b04-4e56-81cc-a632b53293b6\") " pod="openstack/glance-db-sync-wbxqj" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.877031 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6gxs\" (UniqueName: \"kubernetes.io/projected/80a3aebf-6b04-4e56-81cc-a632b53293b6-kube-api-access-p6gxs\") pod \"glance-db-sync-wbxqj\" (UID: \"80a3aebf-6b04-4e56-81cc-a632b53293b6\") " pod="openstack/glance-db-sync-wbxqj" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.889332 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-fmfsh" event={"ID":"033734aa-97c9-498f-8bae-992507b79f71","Type":"ContainerDied","Data":"af7c5ec4bda9c19df9cfaa4dea30203887db49dee36fc652d0f6798631367647"} Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.889377 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af7c5ec4bda9c19df9cfaa4dea30203887db49dee36fc652d0f6798631367647" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.889465 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-fmfsh" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.902687 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65e6-account-create-update-znnqb" event={"ID":"1ec0f64d-691b-4433-a92b-ae6ab0ae4971","Type":"ContainerDied","Data":"e3814a4f6d3b506baf8d5ddfde1c7424e7afa74f1bab2f5b38f16974c82a14ed"} Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.903246 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3814a4f6d3b506baf8d5ddfde1c7424e7afa74f1bab2f5b38f16974c82a14ed" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.903035 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-65e6-account-create-update-znnqb" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.941062 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b025-account-create-update-cxmdd" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.993277 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b025-account-create-update-cxmdd" event={"ID":"373c7c35-5998-40d4-b40b-5024d81e5567","Type":"ContainerDied","Data":"d37c75642aed323a19fc569cba2cfadc1c31c83a932ed4466444c6240731dc17"} Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.993319 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d37c75642aed323a19fc569cba2cfadc1c31c83a932ed4466444c6240731dc17" Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.993335 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-z775c"] Dec 09 10:09:45 crc kubenswrapper[4824]: I1209 10:09:45.993559 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-z775c" podUID="ddc24cb1-2724-40b8-bfc9-81474d67d928" containerName="dnsmasq-dns" containerID="cri-o://39ff5a2a98843182a0574523530ac8dc4f9f7a42fe1e1ed590366daa26ed16aa" gracePeriod=10 Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.196811 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-wbxqj" Dec 09 10:09:46 crc kubenswrapper[4824]: E1209 10:09:46.262255 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721e1031_d80c_406a_ad29_17879d58c3c3.slice/crio-88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c1822c_1257_4d8c_9a17_d921929000fa.slice/crio-conmon-b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod033734aa_97c9_498f_8bae_992507b79f71.slice/crio-af7c5ec4bda9c19df9cfaa4dea30203887db49dee36fc652d0f6798631367647\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod373c7c35_5998_40d4_b40b_5024d81e5567.slice/crio-d37c75642aed323a19fc569cba2cfadc1c31c83a932ed4466444c6240731dc17\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c1822c_1257_4d8c_9a17_d921929000fa.slice/crio-b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721e1031_d80c_406a_ad29_17879d58c3c3.slice/crio-conmon-88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857.scope\": RecentStats: unable to find data in memory cache]" Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.409844 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-985bt-config-x2qzv"] Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.424543 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-985bt-config-x2qzv"] Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.643922 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-985bt-config-tb42c"] Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.645998 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-985bt-config-tb42c" Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.649569 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.672866 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-985bt-config-tb42c"] Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.811606 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-var-run-ovn\") pod \"ovn-controller-985bt-config-tb42c\" (UID: \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\") " pod="openstack/ovn-controller-985bt-config-tb42c" Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.811683 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-var-log-ovn\") pod \"ovn-controller-985bt-config-tb42c\" (UID: \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\") " pod="openstack/ovn-controller-985bt-config-tb42c" Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.812350 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snqxz\" (UniqueName: \"kubernetes.io/projected/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-kube-api-access-snqxz\") pod \"ovn-controller-985bt-config-tb42c\" (UID: \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\") " pod="openstack/ovn-controller-985bt-config-tb42c" Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.812624 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-additional-scripts\") pod \"ovn-controller-985bt-config-tb42c\" (UID: \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\") " pod="openstack/ovn-controller-985bt-config-tb42c" Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.812662 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-var-run\") pod \"ovn-controller-985bt-config-tb42c\" (UID: \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\") " pod="openstack/ovn-controller-985bt-config-tb42c" Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.812738 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-scripts\") pod \"ovn-controller-985bt-config-tb42c\" (UID: \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\") " pod="openstack/ovn-controller-985bt-config-tb42c" Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.915194 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-var-run-ovn\") pod \"ovn-controller-985bt-config-tb42c\" (UID: \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\") " pod="openstack/ovn-controller-985bt-config-tb42c" Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.915257 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-var-log-ovn\") pod \"ovn-controller-985bt-config-tb42c\" (UID: \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\") " pod="openstack/ovn-controller-985bt-config-tb42c" Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.915330 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snqxz\" (UniqueName: \"kubernetes.io/projected/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-kube-api-access-snqxz\") pod \"ovn-controller-985bt-config-tb42c\" (UID: \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\") " pod="openstack/ovn-controller-985bt-config-tb42c" Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.915372 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-additional-scripts\") pod \"ovn-controller-985bt-config-tb42c\" (UID: \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\") " pod="openstack/ovn-controller-985bt-config-tb42c" Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.915389 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-var-run\") pod \"ovn-controller-985bt-config-tb42c\" (UID: \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\") " pod="openstack/ovn-controller-985bt-config-tb42c" Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.915424 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-scripts\") pod \"ovn-controller-985bt-config-tb42c\" (UID: \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\") " pod="openstack/ovn-controller-985bt-config-tb42c" Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.916404 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-var-log-ovn\") pod \"ovn-controller-985bt-config-tb42c\" (UID: \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\") " pod="openstack/ovn-controller-985bt-config-tb42c" Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.916481 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-var-run-ovn\") pod \"ovn-controller-985bt-config-tb42c\" (UID: \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\") " pod="openstack/ovn-controller-985bt-config-tb42c" Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.916767 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-var-run\") pod \"ovn-controller-985bt-config-tb42c\" (UID: \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\") " pod="openstack/ovn-controller-985bt-config-tb42c" Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.917216 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-additional-scripts\") pod \"ovn-controller-985bt-config-tb42c\" (UID: \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\") " pod="openstack/ovn-controller-985bt-config-tb42c" Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.917875 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-scripts\") pod \"ovn-controller-985bt-config-tb42c\" (UID: \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\") " pod="openstack/ovn-controller-985bt-config-tb42c" Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.944837 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snqxz\" (UniqueName: \"kubernetes.io/projected/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-kube-api-access-snqxz\") pod \"ovn-controller-985bt-config-tb42c\" (UID: \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\") " pod="openstack/ovn-controller-985bt-config-tb42c" Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.963085 4824 generic.go:334] "Generic (PLEG): container finished" podID="ddc24cb1-2724-40b8-bfc9-81474d67d928" containerID="39ff5a2a98843182a0574523530ac8dc4f9f7a42fe1e1ed590366daa26ed16aa" exitCode=0 Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.963138 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-z775c" event={"ID":"ddc24cb1-2724-40b8-bfc9-81474d67d928","Type":"ContainerDied","Data":"39ff5a2a98843182a0574523530ac8dc4f9f7a42fe1e1ed590366daa26ed16aa"} Dec 09 10:09:46 crc kubenswrapper[4824]: I1209 10:09:46.979182 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-985bt-config-tb42c" Dec 09 10:09:47 crc kubenswrapper[4824]: I1209 10:09:47.134066 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:09:47 crc kubenswrapper[4824]: I1209 10:09:47.199208 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-fwd2s"] Dec 09 10:09:47 crc kubenswrapper[4824]: I1209 10:09:47.201955 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-fwd2s" Dec 09 10:09:47 crc kubenswrapper[4824]: I1209 10:09:47.229702 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-fwd2s"] Dec 09 10:09:47 crc kubenswrapper[4824]: I1209 10:09:47.332004 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/108f0080-5222-4e67-8cef-6911d8ae4ced-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-fwd2s\" (UID: \"108f0080-5222-4e67-8cef-6911d8ae4ced\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-fwd2s" Dec 09 10:09:47 crc kubenswrapper[4824]: I1209 10:09:47.332554 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsmc4\" (UniqueName: \"kubernetes.io/projected/108f0080-5222-4e67-8cef-6911d8ae4ced-kube-api-access-vsmc4\") pod \"mysqld-exporter-openstack-cell1-db-create-fwd2s\" (UID: \"108f0080-5222-4e67-8cef-6911d8ae4ced\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-fwd2s" Dec 09 10:09:47 crc kubenswrapper[4824]: I1209 10:09:47.416547 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-6387-account-create-update-62lwv"] Dec 09 10:09:47 crc kubenswrapper[4824]: I1209 10:09:47.420403 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-6387-account-create-update-62lwv" Dec 09 10:09:47 crc kubenswrapper[4824]: I1209 10:09:47.423063 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-cell1-db-secret" Dec 09 10:09:47 crc kubenswrapper[4824]: I1209 10:09:47.437168 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/108f0080-5222-4e67-8cef-6911d8ae4ced-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-fwd2s\" (UID: \"108f0080-5222-4e67-8cef-6911d8ae4ced\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-fwd2s" Dec 09 10:09:47 crc kubenswrapper[4824]: I1209 10:09:47.437240 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsmc4\" (UniqueName: \"kubernetes.io/projected/108f0080-5222-4e67-8cef-6911d8ae4ced-kube-api-access-vsmc4\") pod \"mysqld-exporter-openstack-cell1-db-create-fwd2s\" (UID: \"108f0080-5222-4e67-8cef-6911d8ae4ced\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-fwd2s" Dec 09 10:09:47 crc kubenswrapper[4824]: I1209 10:09:47.442029 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/108f0080-5222-4e67-8cef-6911d8ae4ced-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-fwd2s\" (UID: \"108f0080-5222-4e67-8cef-6911d8ae4ced\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-fwd2s" Dec 09 10:09:47 crc kubenswrapper[4824]: I1209 10:09:47.446001 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-6387-account-create-update-62lwv"] Dec 09 10:09:47 crc kubenswrapper[4824]: I1209 10:09:47.479001 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsmc4\" (UniqueName: \"kubernetes.io/projected/108f0080-5222-4e67-8cef-6911d8ae4ced-kube-api-access-vsmc4\") pod \"mysqld-exporter-openstack-cell1-db-create-fwd2s\" (UID: \"108f0080-5222-4e67-8cef-6911d8ae4ced\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-fwd2s" Dec 09 10:09:47 crc kubenswrapper[4824]: I1209 10:09:47.531142 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-fwd2s" Dec 09 10:09:47 crc kubenswrapper[4824]: I1209 10:09:47.539448 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49a2f3f0-02e4-4ed8-93ed-04911062274a-operator-scripts\") pod \"mysqld-exporter-6387-account-create-update-62lwv\" (UID: \"49a2f3f0-02e4-4ed8-93ed-04911062274a\") " pod="openstack/mysqld-exporter-6387-account-create-update-62lwv" Dec 09 10:09:47 crc kubenswrapper[4824]: I1209 10:09:47.539559 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc62h\" (UniqueName: \"kubernetes.io/projected/49a2f3f0-02e4-4ed8-93ed-04911062274a-kube-api-access-tc62h\") pod \"mysqld-exporter-6387-account-create-update-62lwv\" (UID: \"49a2f3f0-02e4-4ed8-93ed-04911062274a\") " pod="openstack/mysqld-exporter-6387-account-create-update-62lwv" Dec 09 10:09:47 crc kubenswrapper[4824]: I1209 10:09:47.642295 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc62h\" (UniqueName: \"kubernetes.io/projected/49a2f3f0-02e4-4ed8-93ed-04911062274a-kube-api-access-tc62h\") pod \"mysqld-exporter-6387-account-create-update-62lwv\" (UID: \"49a2f3f0-02e4-4ed8-93ed-04911062274a\") " pod="openstack/mysqld-exporter-6387-account-create-update-62lwv" Dec 09 10:09:47 crc kubenswrapper[4824]: I1209 10:09:47.642575 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49a2f3f0-02e4-4ed8-93ed-04911062274a-operator-scripts\") pod \"mysqld-exporter-6387-account-create-update-62lwv\" (UID: \"49a2f3f0-02e4-4ed8-93ed-04911062274a\") " pod="openstack/mysqld-exporter-6387-account-create-update-62lwv" Dec 09 10:09:47 crc kubenswrapper[4824]: I1209 10:09:47.643494 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49a2f3f0-02e4-4ed8-93ed-04911062274a-operator-scripts\") pod \"mysqld-exporter-6387-account-create-update-62lwv\" (UID: \"49a2f3f0-02e4-4ed8-93ed-04911062274a\") " pod="openstack/mysqld-exporter-6387-account-create-update-62lwv" Dec 09 10:09:47 crc kubenswrapper[4824]: I1209 10:09:47.663026 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc62h\" (UniqueName: \"kubernetes.io/projected/49a2f3f0-02e4-4ed8-93ed-04911062274a-kube-api-access-tc62h\") pod \"mysqld-exporter-6387-account-create-update-62lwv\" (UID: \"49a2f3f0-02e4-4ed8-93ed-04911062274a\") " pod="openstack/mysqld-exporter-6387-account-create-update-62lwv" Dec 09 10:09:47 crc kubenswrapper[4824]: I1209 10:09:47.751338 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-6387-account-create-update-62lwv" Dec 09 10:09:47 crc kubenswrapper[4824]: I1209 10:09:47.927302 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79e597f5-64c0-4250-b3d4-69ac3db9cb69" path="/var/lib/kubelet/pods/79e597f5-64c0-4250-b3d4-69ac3db9cb69/volumes" Dec 09 10:09:48 crc kubenswrapper[4824]: E1209 10:09:48.107703 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721e1031_d80c_406a_ad29_17879d58c3c3.slice/crio-conmon-88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721e1031_d80c_406a_ad29_17879d58c3c3.slice/crio-88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c1822c_1257_4d8c_9a17_d921929000fa.slice/crio-b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c1822c_1257_4d8c_9a17_d921929000fa.slice/crio-conmon-b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c.scope\": RecentStats: unable to find data in memory cache]" Dec 09 10:09:48 crc kubenswrapper[4824]: E1209 10:09:48.109123 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721e1031_d80c_406a_ad29_17879d58c3c3.slice/crio-conmon-88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c1822c_1257_4d8c_9a17_d921929000fa.slice/crio-b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721e1031_d80c_406a_ad29_17879d58c3c3.slice/crio-88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c1822c_1257_4d8c_9a17_d921929000fa.slice/crio-conmon-b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c.scope\": RecentStats: unable to find data in memory cache]" Dec 09 10:09:49 crc kubenswrapper[4824]: I1209 10:09:49.459675 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-86db49b7ff-z775c" podUID="ddc24cb1-2724-40b8-bfc9-81474d67d928" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.146:5353: connect: connection refused" Dec 09 10:09:50 crc kubenswrapper[4824]: I1209 10:09:50.931734 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-etc-swift\") pod \"swift-storage-0\" (UID: \"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f\") " pod="openstack/swift-storage-0" Dec 09 10:09:50 crc kubenswrapper[4824]: E1209 10:09:50.932030 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 09 10:09:50 crc kubenswrapper[4824]: E1209 10:09:50.932171 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 09 10:09:50 crc kubenswrapper[4824]: E1209 10:09:50.932251 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-etc-swift podName:66fd04f1-a0d9-4295-ac42-e5b20acc7c1f nodeName:}" failed. No retries permitted until 2025-12-09 10:10:06.932229659 +0000 UTC m=+1363.266734326 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-etc-swift") pod "swift-storage-0" (UID: "66fd04f1-a0d9-4295-ac42-e5b20acc7c1f") : configmap "swift-ring-files" not found Dec 09 10:09:51 crc kubenswrapper[4824]: I1209 10:09:51.835128 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-z775c" Dec 09 10:09:51 crc kubenswrapper[4824]: I1209 10:09:51.856004 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2n7h2\" (UniqueName: \"kubernetes.io/projected/ddc24cb1-2724-40b8-bfc9-81474d67d928-kube-api-access-2n7h2\") pod \"ddc24cb1-2724-40b8-bfc9-81474d67d928\" (UID: \"ddc24cb1-2724-40b8-bfc9-81474d67d928\") " Dec 09 10:09:51 crc kubenswrapper[4824]: I1209 10:09:51.856469 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ddc24cb1-2724-40b8-bfc9-81474d67d928-ovsdbserver-sb\") pod \"ddc24cb1-2724-40b8-bfc9-81474d67d928\" (UID: \"ddc24cb1-2724-40b8-bfc9-81474d67d928\") " Dec 09 10:09:51 crc kubenswrapper[4824]: I1209 10:09:51.876337 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ddc24cb1-2724-40b8-bfc9-81474d67d928-ovsdbserver-nb\") pod \"ddc24cb1-2724-40b8-bfc9-81474d67d928\" (UID: \"ddc24cb1-2724-40b8-bfc9-81474d67d928\") " Dec 09 10:09:51 crc kubenswrapper[4824]: I1209 10:09:51.876384 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ddc24cb1-2724-40b8-bfc9-81474d67d928-dns-svc\") pod \"ddc24cb1-2724-40b8-bfc9-81474d67d928\" (UID: \"ddc24cb1-2724-40b8-bfc9-81474d67d928\") " Dec 09 10:09:51 crc kubenswrapper[4824]: I1209 10:09:51.876445 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddc24cb1-2724-40b8-bfc9-81474d67d928-config\") pod \"ddc24cb1-2724-40b8-bfc9-81474d67d928\" (UID: \"ddc24cb1-2724-40b8-bfc9-81474d67d928\") " Dec 09 10:09:51 crc kubenswrapper[4824]: I1209 10:09:51.878888 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddc24cb1-2724-40b8-bfc9-81474d67d928-kube-api-access-2n7h2" (OuterVolumeSpecName: "kube-api-access-2n7h2") pod "ddc24cb1-2724-40b8-bfc9-81474d67d928" (UID: "ddc24cb1-2724-40b8-bfc9-81474d67d928"). InnerVolumeSpecName "kube-api-access-2n7h2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:09:51 crc kubenswrapper[4824]: I1209 10:09:51.882260 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2n7h2\" (UniqueName: \"kubernetes.io/projected/ddc24cb1-2724-40b8-bfc9-81474d67d928-kube-api-access-2n7h2\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:52 crc kubenswrapper[4824]: I1209 10:09:52.027514 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-z775c" event={"ID":"ddc24cb1-2724-40b8-bfc9-81474d67d928","Type":"ContainerDied","Data":"8c48407177b665d20502ea8c60de2bbdab77c7310c46231bab2151641bfb905a"} Dec 09 10:09:52 crc kubenswrapper[4824]: I1209 10:09:52.027574 4824 scope.go:117] "RemoveContainer" containerID="39ff5a2a98843182a0574523530ac8dc4f9f7a42fe1e1ed590366daa26ed16aa" Dec 09 10:09:52 crc kubenswrapper[4824]: I1209 10:09:52.028045 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-z775c" Dec 09 10:09:52 crc kubenswrapper[4824]: I1209 10:09:52.059831 4824 scope.go:117] "RemoveContainer" containerID="31ebf1b50662f71b55fc3e7f1aa08e494f7f99302d8044c600a3e27be0d119a6" Dec 09 10:09:52 crc kubenswrapper[4824]: I1209 10:09:52.082585 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ddc24cb1-2724-40b8-bfc9-81474d67d928-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ddc24cb1-2724-40b8-bfc9-81474d67d928" (UID: "ddc24cb1-2724-40b8-bfc9-81474d67d928"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:52 crc kubenswrapper[4824]: I1209 10:09:52.088509 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ddc24cb1-2724-40b8-bfc9-81474d67d928-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ddc24cb1-2724-40b8-bfc9-81474d67d928" (UID: "ddc24cb1-2724-40b8-bfc9-81474d67d928"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:52 crc kubenswrapper[4824]: I1209 10:09:52.105296 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ddc24cb1-2724-40b8-bfc9-81474d67d928-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:52 crc kubenswrapper[4824]: I1209 10:09:52.105380 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ddc24cb1-2724-40b8-bfc9-81474d67d928-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:52 crc kubenswrapper[4824]: I1209 10:09:52.106374 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ddc24cb1-2724-40b8-bfc9-81474d67d928-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ddc24cb1-2724-40b8-bfc9-81474d67d928" (UID: "ddc24cb1-2724-40b8-bfc9-81474d67d928"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:52 crc kubenswrapper[4824]: I1209 10:09:52.130007 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ddc24cb1-2724-40b8-bfc9-81474d67d928-config" (OuterVolumeSpecName: "config") pod "ddc24cb1-2724-40b8-bfc9-81474d67d928" (UID: "ddc24cb1-2724-40b8-bfc9-81474d67d928"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:52 crc kubenswrapper[4824]: I1209 10:09:52.207457 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ddc24cb1-2724-40b8-bfc9-81474d67d928-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:52 crc kubenswrapper[4824]: I1209 10:09:52.207545 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddc24cb1-2724-40b8-bfc9-81474d67d928-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:52 crc kubenswrapper[4824]: W1209 10:09:52.216698 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod108f0080_5222_4e67_8cef_6911d8ae4ced.slice/crio-5a4a0f9e30bf99cac468a5268d1b3ebf637db5e072885890a168c85d8d9a95ec WatchSource:0}: Error finding container 5a4a0f9e30bf99cac468a5268d1b3ebf637db5e072885890a168c85d8d9a95ec: Status 404 returned error can't find the container with id 5a4a0f9e30bf99cac468a5268d1b3ebf637db5e072885890a168c85d8d9a95ec Dec 09 10:09:52 crc kubenswrapper[4824]: I1209 10:09:52.218497 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-fwd2s"] Dec 09 10:09:52 crc kubenswrapper[4824]: I1209 10:09:52.306661 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-6387-account-create-update-62lwv"] Dec 09 10:09:52 crc kubenswrapper[4824]: W1209 10:09:52.350185 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49a2f3f0_02e4_4ed8_93ed_04911062274a.slice/crio-4a2f72843fc9e32ca2dd4357e7b87b5993509fe605c71e438ae4cdb824db95fd WatchSource:0}: Error finding container 4a2f72843fc9e32ca2dd4357e7b87b5993509fe605c71e438ae4cdb824db95fd: Status 404 returned error can't find the container with id 4a2f72843fc9e32ca2dd4357e7b87b5993509fe605c71e438ae4cdb824db95fd Dec 09 10:09:52 crc kubenswrapper[4824]: I1209 10:09:52.412135 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-z775c"] Dec 09 10:09:52 crc kubenswrapper[4824]: I1209 10:09:52.442367 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-z775c"] Dec 09 10:09:52 crc kubenswrapper[4824]: I1209 10:09:52.532799 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-985bt-config-tb42c"] Dec 09 10:09:52 crc kubenswrapper[4824]: I1209 10:09:52.593192 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-wbxqj"] Dec 09 10:09:53 crc kubenswrapper[4824]: I1209 10:09:53.049341 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-wbxqj" event={"ID":"80a3aebf-6b04-4e56-81cc-a632b53293b6","Type":"ContainerStarted","Data":"d9f1f3112821b81ab615cb06293cd5943dbedb7fe1be424fe7a4c30c50c7c103"} Dec 09 10:09:53 crc kubenswrapper[4824]: I1209 10:09:53.051805 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-cf5wn" event={"ID":"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c","Type":"ContainerStarted","Data":"2ecc1183fb08c8a54f6f0fac74be0e24ffac68b3d868de43ef30256b3240a6ff"} Dec 09 10:09:53 crc kubenswrapper[4824]: I1209 10:09:53.056845 4824 generic.go:334] "Generic (PLEG): container finished" podID="49a2f3f0-02e4-4ed8-93ed-04911062274a" containerID="962c26d0568bf98785f55af084fd2975bb8811ab3e95b72ecb01db06de61f1c4" exitCode=0 Dec 09 10:09:53 crc kubenswrapper[4824]: I1209 10:09:53.056951 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-6387-account-create-update-62lwv" event={"ID":"49a2f3f0-02e4-4ed8-93ed-04911062274a","Type":"ContainerDied","Data":"962c26d0568bf98785f55af084fd2975bb8811ab3e95b72ecb01db06de61f1c4"} Dec 09 10:09:53 crc kubenswrapper[4824]: I1209 10:09:53.056999 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-6387-account-create-update-62lwv" event={"ID":"49a2f3f0-02e4-4ed8-93ed-04911062274a","Type":"ContainerStarted","Data":"4a2f72843fc9e32ca2dd4357e7b87b5993509fe605c71e438ae4cdb824db95fd"} Dec 09 10:09:53 crc kubenswrapper[4824]: I1209 10:09:53.063937 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8edb90f3-f84b-4700-993f-21ffa950b620","Type":"ContainerStarted","Data":"065fe900e020f72eec27757202af458fb43910dff8bc9cca1e8b163284ff096f"} Dec 09 10:09:53 crc kubenswrapper[4824]: I1209 10:09:53.071977 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-cf5wn" podStartSLOduration=11.56879684 podStartE2EDuration="26.071952935s" podCreationTimestamp="2025-12-09 10:09:27 +0000 UTC" firstStartedPulling="2025-12-09 10:09:37.14853717 +0000 UTC m=+1333.483041837" lastFinishedPulling="2025-12-09 10:09:51.651693265 +0000 UTC m=+1347.986197932" observedRunningTime="2025-12-09 10:09:53.069885138 +0000 UTC m=+1349.404389825" watchObservedRunningTime="2025-12-09 10:09:53.071952935 +0000 UTC m=+1349.406457622" Dec 09 10:09:53 crc kubenswrapper[4824]: I1209 10:09:53.073457 4824 generic.go:334] "Generic (PLEG): container finished" podID="108f0080-5222-4e67-8cef-6911d8ae4ced" containerID="1478290470c9d624da9f9cfc256d4304430790b207cbe3f1255acb46080432f6" exitCode=0 Dec 09 10:09:53 crc kubenswrapper[4824]: I1209 10:09:53.073552 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-fwd2s" event={"ID":"108f0080-5222-4e67-8cef-6911d8ae4ced","Type":"ContainerDied","Data":"1478290470c9d624da9f9cfc256d4304430790b207cbe3f1255acb46080432f6"} Dec 09 10:09:53 crc kubenswrapper[4824]: I1209 10:09:53.073597 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-fwd2s" event={"ID":"108f0080-5222-4e67-8cef-6911d8ae4ced","Type":"ContainerStarted","Data":"5a4a0f9e30bf99cac468a5268d1b3ebf637db5e072885890a168c85d8d9a95ec"} Dec 09 10:09:53 crc kubenswrapper[4824]: I1209 10:09:53.085737 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-985bt-config-tb42c" event={"ID":"9a1f6de2-d358-49f2-b564-d0a9e16cb73e","Type":"ContainerStarted","Data":"b1ce36f3c2bba4bc9b2767a9ae4377d6ad2af28c7b6ba4c6368961cc6fe41705"} Dec 09 10:09:53 crc kubenswrapper[4824]: I1209 10:09:53.085846 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-985bt-config-tb42c" event={"ID":"9a1f6de2-d358-49f2-b564-d0a9e16cb73e","Type":"ContainerStarted","Data":"dd063833affc38ab2828d72887c6490f62771aedbc609b4375bb0a0b970f529b"} Dec 09 10:09:53 crc kubenswrapper[4824]: I1209 10:09:53.132135 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-985bt-config-tb42c" podStartSLOduration=7.132117151 podStartE2EDuration="7.132117151s" podCreationTimestamp="2025-12-09 10:09:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:09:53.127982391 +0000 UTC m=+1349.462487068" watchObservedRunningTime="2025-12-09 10:09:53.132117151 +0000 UTC m=+1349.466621818" Dec 09 10:09:53 crc kubenswrapper[4824]: E1209 10:09:53.494396 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721e1031_d80c_406a_ad29_17879d58c3c3.slice/crio-88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721e1031_d80c_406a_ad29_17879d58c3c3.slice/crio-conmon-88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c1822c_1257_4d8c_9a17_d921929000fa.slice/crio-conmon-b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c1822c_1257_4d8c_9a17_d921929000fa.slice/crio-b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c.scope\": RecentStats: unable to find data in memory cache]" Dec 09 10:09:53 crc kubenswrapper[4824]: I1209 10:09:53.924506 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddc24cb1-2724-40b8-bfc9-81474d67d928" path="/var/lib/kubelet/pods/ddc24cb1-2724-40b8-bfc9-81474d67d928/volumes" Dec 09 10:09:54 crc kubenswrapper[4824]: I1209 10:09:54.100592 4824 generic.go:334] "Generic (PLEG): container finished" podID="9a1f6de2-d358-49f2-b564-d0a9e16cb73e" containerID="b1ce36f3c2bba4bc9b2767a9ae4377d6ad2af28c7b6ba4c6368961cc6fe41705" exitCode=0 Dec 09 10:09:54 crc kubenswrapper[4824]: I1209 10:09:54.100707 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-985bt-config-tb42c" event={"ID":"9a1f6de2-d358-49f2-b564-d0a9e16cb73e","Type":"ContainerDied","Data":"b1ce36f3c2bba4bc9b2767a9ae4377d6ad2af28c7b6ba4c6368961cc6fe41705"} Dec 09 10:09:54 crc kubenswrapper[4824]: I1209 10:09:54.689934 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-fwd2s" Dec 09 10:09:54 crc kubenswrapper[4824]: I1209 10:09:54.698024 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-6387-account-create-update-62lwv" Dec 09 10:09:54 crc kubenswrapper[4824]: I1209 10:09:54.782284 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/108f0080-5222-4e67-8cef-6911d8ae4ced-operator-scripts\") pod \"108f0080-5222-4e67-8cef-6911d8ae4ced\" (UID: \"108f0080-5222-4e67-8cef-6911d8ae4ced\") " Dec 09 10:09:54 crc kubenswrapper[4824]: I1209 10:09:54.782328 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49a2f3f0-02e4-4ed8-93ed-04911062274a-operator-scripts\") pod \"49a2f3f0-02e4-4ed8-93ed-04911062274a\" (UID: \"49a2f3f0-02e4-4ed8-93ed-04911062274a\") " Dec 09 10:09:54 crc kubenswrapper[4824]: I1209 10:09:54.782465 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsmc4\" (UniqueName: \"kubernetes.io/projected/108f0080-5222-4e67-8cef-6911d8ae4ced-kube-api-access-vsmc4\") pod \"108f0080-5222-4e67-8cef-6911d8ae4ced\" (UID: \"108f0080-5222-4e67-8cef-6911d8ae4ced\") " Dec 09 10:09:54 crc kubenswrapper[4824]: I1209 10:09:54.782635 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tc62h\" (UniqueName: \"kubernetes.io/projected/49a2f3f0-02e4-4ed8-93ed-04911062274a-kube-api-access-tc62h\") pod \"49a2f3f0-02e4-4ed8-93ed-04911062274a\" (UID: \"49a2f3f0-02e4-4ed8-93ed-04911062274a\") " Dec 09 10:09:54 crc kubenswrapper[4824]: I1209 10:09:54.783266 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49a2f3f0-02e4-4ed8-93ed-04911062274a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "49a2f3f0-02e4-4ed8-93ed-04911062274a" (UID: "49a2f3f0-02e4-4ed8-93ed-04911062274a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:54 crc kubenswrapper[4824]: I1209 10:09:54.783261 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/108f0080-5222-4e67-8cef-6911d8ae4ced-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "108f0080-5222-4e67-8cef-6911d8ae4ced" (UID: "108f0080-5222-4e67-8cef-6911d8ae4ced"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:54 crc kubenswrapper[4824]: I1209 10:09:54.885903 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/108f0080-5222-4e67-8cef-6911d8ae4ced-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:54 crc kubenswrapper[4824]: I1209 10:09:54.885949 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49a2f3f0-02e4-4ed8-93ed-04911062274a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.044041 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49a2f3f0-02e4-4ed8-93ed-04911062274a-kube-api-access-tc62h" (OuterVolumeSpecName: "kube-api-access-tc62h") pod "49a2f3f0-02e4-4ed8-93ed-04911062274a" (UID: "49a2f3f0-02e4-4ed8-93ed-04911062274a"). InnerVolumeSpecName "kube-api-access-tc62h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.044232 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/108f0080-5222-4e67-8cef-6911d8ae4ced-kube-api-access-vsmc4" (OuterVolumeSpecName: "kube-api-access-vsmc4") pod "108f0080-5222-4e67-8cef-6911d8ae4ced" (UID: "108f0080-5222-4e67-8cef-6911d8ae4ced"). InnerVolumeSpecName "kube-api-access-vsmc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.094435 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsmc4\" (UniqueName: \"kubernetes.io/projected/108f0080-5222-4e67-8cef-6911d8ae4ced-kube-api-access-vsmc4\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.094486 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tc62h\" (UniqueName: \"kubernetes.io/projected/49a2f3f0-02e4-4ed8-93ed-04911062274a-kube-api-access-tc62h\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.112268 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-fwd2s" event={"ID":"108f0080-5222-4e67-8cef-6911d8ae4ced","Type":"ContainerDied","Data":"5a4a0f9e30bf99cac468a5268d1b3ebf637db5e072885890a168c85d8d9a95ec"} Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.112313 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a4a0f9e30bf99cac468a5268d1b3ebf637db5e072885890a168c85d8d9a95ec" Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.112373 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-fwd2s" Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.117448 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-6387-account-create-update-62lwv" Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.120340 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-6387-account-create-update-62lwv" event={"ID":"49a2f3f0-02e4-4ed8-93ed-04911062274a","Type":"ContainerDied","Data":"4a2f72843fc9e32ca2dd4357e7b87b5993509fe605c71e438ae4cdb824db95fd"} Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.120389 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a2f72843fc9e32ca2dd4357e7b87b5993509fe605c71e438ae4cdb824db95fd" Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.497177 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-985bt-config-tb42c" Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.592368 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-985bt-config-tb42c"] Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.601900 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-985bt-config-tb42c"] Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.607974 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-var-run-ovn\") pod \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\" (UID: \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\") " Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.608113 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-var-log-ovn\") pod \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\" (UID: \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\") " Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.608223 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snqxz\" (UniqueName: \"kubernetes.io/projected/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-kube-api-access-snqxz\") pod \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\" (UID: \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\") " Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.608221 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "9a1f6de2-d358-49f2-b564-d0a9e16cb73e" (UID: "9a1f6de2-d358-49f2-b564-d0a9e16cb73e"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.608251 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-scripts\") pod \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\" (UID: \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\") " Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.608282 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "9a1f6de2-d358-49f2-b564-d0a9e16cb73e" (UID: "9a1f6de2-d358-49f2-b564-d0a9e16cb73e"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.608300 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-var-run\") pod \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\" (UID: \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\") " Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.608324 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-additional-scripts\") pod \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\" (UID: \"9a1f6de2-d358-49f2-b564-d0a9e16cb73e\") " Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.608823 4824 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.608838 4824 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.609591 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "9a1f6de2-d358-49f2-b564-d0a9e16cb73e" (UID: "9a1f6de2-d358-49f2-b564-d0a9e16cb73e"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.609951 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-var-run" (OuterVolumeSpecName: "var-run") pod "9a1f6de2-d358-49f2-b564-d0a9e16cb73e" (UID: "9a1f6de2-d358-49f2-b564-d0a9e16cb73e"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.610006 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-scripts" (OuterVolumeSpecName: "scripts") pod "9a1f6de2-d358-49f2-b564-d0a9e16cb73e" (UID: "9a1f6de2-d358-49f2-b564-d0a9e16cb73e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.614376 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-kube-api-access-snqxz" (OuterVolumeSpecName: "kube-api-access-snqxz") pod "9a1f6de2-d358-49f2-b564-d0a9e16cb73e" (UID: "9a1f6de2-d358-49f2-b564-d0a9e16cb73e"). InnerVolumeSpecName "kube-api-access-snqxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.726559 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snqxz\" (UniqueName: \"kubernetes.io/projected/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-kube-api-access-snqxz\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.730519 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.730564 4824 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-var-run\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.730577 4824 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9a1f6de2-d358-49f2-b564-d0a9e16cb73e-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:09:55 crc kubenswrapper[4824]: I1209 10:09:55.940956 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a1f6de2-d358-49f2-b564-d0a9e16cb73e" path="/var/lib/kubelet/pods/9a1f6de2-d358-49f2-b564-d0a9e16cb73e/volumes" Dec 09 10:09:56 crc kubenswrapper[4824]: I1209 10:09:56.132415 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8edb90f3-f84b-4700-993f-21ffa950b620","Type":"ContainerStarted","Data":"3bcac543b2d89d00cda20308eec14bbc4383f0f9d93d98669cd2beae0742c17f"} Dec 09 10:09:56 crc kubenswrapper[4824]: I1209 10:09:56.134711 4824 scope.go:117] "RemoveContainer" containerID="b1ce36f3c2bba4bc9b2767a9ae4377d6ad2af28c7b6ba4c6368961cc6fe41705" Dec 09 10:09:56 crc kubenswrapper[4824]: I1209 10:09:56.134922 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-985bt-config-tb42c" Dec 09 10:09:56 crc kubenswrapper[4824]: E1209 10:09:56.317549 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c1822c_1257_4d8c_9a17_d921929000fa.slice/crio-conmon-b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721e1031_d80c_406a_ad29_17879d58c3c3.slice/crio-88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721e1031_d80c_406a_ad29_17879d58c3c3.slice/crio-conmon-88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c1822c_1257_4d8c_9a17_d921929000fa.slice/crio-b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c.scope\": RecentStats: unable to find data in memory cache]" Dec 09 10:09:57 crc kubenswrapper[4824]: I1209 10:09:57.627172 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Dec 09 10:09:57 crc kubenswrapper[4824]: E1209 10:09:57.628266 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a1f6de2-d358-49f2-b564-d0a9e16cb73e" containerName="ovn-config" Dec 09 10:09:57 crc kubenswrapper[4824]: I1209 10:09:57.628286 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a1f6de2-d358-49f2-b564-d0a9e16cb73e" containerName="ovn-config" Dec 09 10:09:57 crc kubenswrapper[4824]: E1209 10:09:57.628309 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddc24cb1-2724-40b8-bfc9-81474d67d928" containerName="init" Dec 09 10:09:57 crc kubenswrapper[4824]: I1209 10:09:57.628317 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddc24cb1-2724-40b8-bfc9-81474d67d928" containerName="init" Dec 09 10:09:57 crc kubenswrapper[4824]: E1209 10:09:57.628338 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="108f0080-5222-4e67-8cef-6911d8ae4ced" containerName="mariadb-database-create" Dec 09 10:09:57 crc kubenswrapper[4824]: I1209 10:09:57.628349 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="108f0080-5222-4e67-8cef-6911d8ae4ced" containerName="mariadb-database-create" Dec 09 10:09:57 crc kubenswrapper[4824]: E1209 10:09:57.628373 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49a2f3f0-02e4-4ed8-93ed-04911062274a" containerName="mariadb-account-create-update" Dec 09 10:09:57 crc kubenswrapper[4824]: I1209 10:09:57.628383 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="49a2f3f0-02e4-4ed8-93ed-04911062274a" containerName="mariadb-account-create-update" Dec 09 10:09:57 crc kubenswrapper[4824]: E1209 10:09:57.628420 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddc24cb1-2724-40b8-bfc9-81474d67d928" containerName="dnsmasq-dns" Dec 09 10:09:57 crc kubenswrapper[4824]: I1209 10:09:57.628427 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddc24cb1-2724-40b8-bfc9-81474d67d928" containerName="dnsmasq-dns" Dec 09 10:09:57 crc kubenswrapper[4824]: I1209 10:09:57.628660 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="108f0080-5222-4e67-8cef-6911d8ae4ced" containerName="mariadb-database-create" Dec 09 10:09:57 crc kubenswrapper[4824]: I1209 10:09:57.628676 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a1f6de2-d358-49f2-b564-d0a9e16cb73e" containerName="ovn-config" Dec 09 10:09:57 crc kubenswrapper[4824]: I1209 10:09:57.628694 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddc24cb1-2724-40b8-bfc9-81474d67d928" containerName="dnsmasq-dns" Dec 09 10:09:57 crc kubenswrapper[4824]: I1209 10:09:57.628708 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="49a2f3f0-02e4-4ed8-93ed-04911062274a" containerName="mariadb-account-create-update" Dec 09 10:09:57 crc kubenswrapper[4824]: I1209 10:09:57.630156 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 09 10:09:57 crc kubenswrapper[4824]: I1209 10:09:57.636387 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Dec 09 10:09:57 crc kubenswrapper[4824]: I1209 10:09:57.645289 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 09 10:09:57 crc kubenswrapper[4824]: I1209 10:09:57.694932 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d47a992-6428-4729-8f4d-04af792e688e-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"8d47a992-6428-4729-8f4d-04af792e688e\") " pod="openstack/mysqld-exporter-0" Dec 09 10:09:57 crc kubenswrapper[4824]: I1209 10:09:57.695039 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d47a992-6428-4729-8f4d-04af792e688e-config-data\") pod \"mysqld-exporter-0\" (UID: \"8d47a992-6428-4729-8f4d-04af792e688e\") " pod="openstack/mysqld-exporter-0" Dec 09 10:09:57 crc kubenswrapper[4824]: I1209 10:09:57.695101 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltv52\" (UniqueName: \"kubernetes.io/projected/8d47a992-6428-4729-8f4d-04af792e688e-kube-api-access-ltv52\") pod \"mysqld-exporter-0\" (UID: \"8d47a992-6428-4729-8f4d-04af792e688e\") " pod="openstack/mysqld-exporter-0" Dec 09 10:09:57 crc kubenswrapper[4824]: I1209 10:09:57.797700 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d47a992-6428-4729-8f4d-04af792e688e-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"8d47a992-6428-4729-8f4d-04af792e688e\") " pod="openstack/mysqld-exporter-0" Dec 09 10:09:57 crc kubenswrapper[4824]: I1209 10:09:57.797823 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d47a992-6428-4729-8f4d-04af792e688e-config-data\") pod \"mysqld-exporter-0\" (UID: \"8d47a992-6428-4729-8f4d-04af792e688e\") " pod="openstack/mysqld-exporter-0" Dec 09 10:09:57 crc kubenswrapper[4824]: I1209 10:09:57.797906 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltv52\" (UniqueName: \"kubernetes.io/projected/8d47a992-6428-4729-8f4d-04af792e688e-kube-api-access-ltv52\") pod \"mysqld-exporter-0\" (UID: \"8d47a992-6428-4729-8f4d-04af792e688e\") " pod="openstack/mysqld-exporter-0" Dec 09 10:09:57 crc kubenswrapper[4824]: I1209 10:09:57.806609 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d47a992-6428-4729-8f4d-04af792e688e-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"8d47a992-6428-4729-8f4d-04af792e688e\") " pod="openstack/mysqld-exporter-0" Dec 09 10:09:57 crc kubenswrapper[4824]: I1209 10:09:57.808119 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d47a992-6428-4729-8f4d-04af792e688e-config-data\") pod \"mysqld-exporter-0\" (UID: \"8d47a992-6428-4729-8f4d-04af792e688e\") " pod="openstack/mysqld-exporter-0" Dec 09 10:09:57 crc kubenswrapper[4824]: I1209 10:09:57.819512 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltv52\" (UniqueName: \"kubernetes.io/projected/8d47a992-6428-4729-8f4d-04af792e688e-kube-api-access-ltv52\") pod \"mysqld-exporter-0\" (UID: \"8d47a992-6428-4729-8f4d-04af792e688e\") " pod="openstack/mysqld-exporter-0" Dec 09 10:09:57 crc kubenswrapper[4824]: I1209 10:09:57.985887 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 09 10:10:01 crc kubenswrapper[4824]: I1209 10:10:01.223500 4824 generic.go:334] "Generic (PLEG): container finished" podID="abbc90b3-afb8-4c42-a07d-1d1c38b8d87c" containerID="2ecc1183fb08c8a54f6f0fac74be0e24ffac68b3d868de43ef30256b3240a6ff" exitCode=0 Dec 09 10:10:01 crc kubenswrapper[4824]: I1209 10:10:01.223707 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-cf5wn" event={"ID":"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c","Type":"ContainerDied","Data":"2ecc1183fb08c8a54f6f0fac74be0e24ffac68b3d868de43ef30256b3240a6ff"} Dec 09 10:10:06 crc kubenswrapper[4824]: I1209 10:10:06.286918 4824 generic.go:334] "Generic (PLEG): container finished" podID="cee7a7d4-090b-4db0-abf1-d3f861f8e594" containerID="9340d5745e62248e475ec029542c6a7717fa8bb811727daef75c5eb26017e816" exitCode=0 Dec 09 10:10:06 crc kubenswrapper[4824]: I1209 10:10:06.286999 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"cee7a7d4-090b-4db0-abf1-d3f861f8e594","Type":"ContainerDied","Data":"9340d5745e62248e475ec029542c6a7717fa8bb811727daef75c5eb26017e816"} Dec 09 10:10:06 crc kubenswrapper[4824]: E1209 10:10:06.366239 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c1822c_1257_4d8c_9a17_d921929000fa.slice/crio-b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721e1031_d80c_406a_ad29_17879d58c3c3.slice/crio-88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721e1031_d80c_406a_ad29_17879d58c3c3.slice/crio-conmon-88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c1822c_1257_4d8c_9a17_d921929000fa.slice/crio-conmon-b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c.scope\": RecentStats: unable to find data in memory cache]" Dec 09 10:10:06 crc kubenswrapper[4824]: I1209 10:10:06.968733 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-etc-swift\") pod \"swift-storage-0\" (UID: \"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f\") " pod="openstack/swift-storage-0" Dec 09 10:10:06 crc kubenswrapper[4824]: I1209 10:10:06.980772 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/66fd04f1-a0d9-4295-ac42-e5b20acc7c1f-etc-swift\") pod \"swift-storage-0\" (UID: \"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f\") " pod="openstack/swift-storage-0" Dec 09 10:10:07 crc kubenswrapper[4824]: I1209 10:10:07.130418 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 09 10:10:07 crc kubenswrapper[4824]: I1209 10:10:07.301785 4824 generic.go:334] "Generic (PLEG): container finished" podID="30eb761a-7bff-43a5-adfb-40e90e17e4a6" containerID="24701089b3049252c9f2d6a4217bcb6c2856b5c88e9859477a2fe57787dc088a" exitCode=0 Dec 09 10:10:07 crc kubenswrapper[4824]: I1209 10:10:07.302025 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"30eb761a-7bff-43a5-adfb-40e90e17e4a6","Type":"ContainerDied","Data":"24701089b3049252c9f2d6a4217bcb6c2856b5c88e9859477a2fe57787dc088a"} Dec 09 10:10:08 crc kubenswrapper[4824]: E1209 10:10:08.516119 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c1822c_1257_4d8c_9a17_d921929000fa.slice/crio-conmon-b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721e1031_d80c_406a_ad29_17879d58c3c3.slice/crio-conmon-88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c1822c_1257_4d8c_9a17_d921929000fa.slice/crio-b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721e1031_d80c_406a_ad29_17879d58c3c3.slice/crio-88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857.scope\": RecentStats: unable to find data in memory cache]" Dec 09 10:10:10 crc kubenswrapper[4824]: E1209 10:10:10.270969 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Dec 09 10:10:10 crc kubenswrapper[4824]: E1209 10:10:10.271484 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p6gxs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-wbxqj_openstack(80a3aebf-6b04-4e56-81cc-a632b53293b6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:10:10 crc kubenswrapper[4824]: E1209 10:10:10.273037 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-wbxqj" podUID="80a3aebf-6b04-4e56-81cc-a632b53293b6" Dec 09 10:10:10 crc kubenswrapper[4824]: E1209 10:10:10.390513 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-wbxqj" podUID="80a3aebf-6b04-4e56-81cc-a632b53293b6" Dec 09 10:10:10 crc kubenswrapper[4824]: I1209 10:10:10.606914 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:10:10 crc kubenswrapper[4824]: I1209 10:10:10.789808 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-ring-data-devices\") pod \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " Dec 09 10:10:10 crc kubenswrapper[4824]: I1209 10:10:10.790340 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-scripts\") pod \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " Dec 09 10:10:10 crc kubenswrapper[4824]: I1209 10:10:10.790398 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-dispersionconf\") pod \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " Dec 09 10:10:10 crc kubenswrapper[4824]: I1209 10:10:10.790429 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-swiftconf\") pod \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " Dec 09 10:10:10 crc kubenswrapper[4824]: I1209 10:10:10.790484 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-etc-swift\") pod \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " Dec 09 10:10:10 crc kubenswrapper[4824]: I1209 10:10:10.790553 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7bks\" (UniqueName: \"kubernetes.io/projected/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-kube-api-access-w7bks\") pod \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " Dec 09 10:10:10 crc kubenswrapper[4824]: I1209 10:10:10.790731 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-combined-ca-bundle\") pod \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\" (UID: \"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c\") " Dec 09 10:10:10 crc kubenswrapper[4824]: I1209 10:10:10.795687 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "abbc90b3-afb8-4c42-a07d-1d1c38b8d87c" (UID: "abbc90b3-afb8-4c42-a07d-1d1c38b8d87c"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:10:10 crc kubenswrapper[4824]: I1209 10:10:10.796562 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "abbc90b3-afb8-4c42-a07d-1d1c38b8d87c" (UID: "abbc90b3-afb8-4c42-a07d-1d1c38b8d87c"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:10:10 crc kubenswrapper[4824]: I1209 10:10:10.835034 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "abbc90b3-afb8-4c42-a07d-1d1c38b8d87c" (UID: "abbc90b3-afb8-4c42-a07d-1d1c38b8d87c"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:10:10 crc kubenswrapper[4824]: I1209 10:10:10.835416 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-scripts" (OuterVolumeSpecName: "scripts") pod "abbc90b3-afb8-4c42-a07d-1d1c38b8d87c" (UID: "abbc90b3-afb8-4c42-a07d-1d1c38b8d87c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:10:10 crc kubenswrapper[4824]: I1209 10:10:10.835724 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-kube-api-access-w7bks" (OuterVolumeSpecName: "kube-api-access-w7bks") pod "abbc90b3-afb8-4c42-a07d-1d1c38b8d87c" (UID: "abbc90b3-afb8-4c42-a07d-1d1c38b8d87c"). InnerVolumeSpecName "kube-api-access-w7bks". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:10:10 crc kubenswrapper[4824]: I1209 10:10:10.864651 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "abbc90b3-afb8-4c42-a07d-1d1c38b8d87c" (UID: "abbc90b3-afb8-4c42-a07d-1d1c38b8d87c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:10:10 crc kubenswrapper[4824]: I1209 10:10:10.896506 4824 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:10 crc kubenswrapper[4824]: I1209 10:10:10.896551 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:10 crc kubenswrapper[4824]: I1209 10:10:10.896569 4824 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:10 crc kubenswrapper[4824]: I1209 10:10:10.896582 4824 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:10 crc kubenswrapper[4824]: I1209 10:10:10.896599 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7bks\" (UniqueName: \"kubernetes.io/projected/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-kube-api-access-w7bks\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:10 crc kubenswrapper[4824]: I1209 10:10:10.896613 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:11 crc kubenswrapper[4824]: I1209 10:10:11.041182 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "abbc90b3-afb8-4c42-a07d-1d1c38b8d87c" (UID: "abbc90b3-afb8-4c42-a07d-1d1c38b8d87c"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:10:11 crc kubenswrapper[4824]: I1209 10:10:11.105863 4824 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/abbc90b3-afb8-4c42-a07d-1d1c38b8d87c-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:11 crc kubenswrapper[4824]: I1209 10:10:11.176252 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 09 10:10:11 crc kubenswrapper[4824]: I1209 10:10:11.385319 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8edb90f3-f84b-4700-993f-21ffa950b620","Type":"ContainerStarted","Data":"99db94c0c3cfc2e1ecab46d0a7b0c3c3d0fcc1067806b3f5b0b09fba74941434"} Dec 09 10:10:11 crc kubenswrapper[4824]: I1209 10:10:11.392415 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"8d47a992-6428-4729-8f4d-04af792e688e","Type":"ContainerStarted","Data":"d6dc549978dfaf666b6b09ba5a2997fd6b3a47fb7a0cd29214e4fc85e0c92702"} Dec 09 10:10:11 crc kubenswrapper[4824]: I1209 10:10:11.395545 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-cf5wn" Dec 09 10:10:11 crc kubenswrapper[4824]: I1209 10:10:11.395574 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-cf5wn" event={"ID":"abbc90b3-afb8-4c42-a07d-1d1c38b8d87c","Type":"ContainerDied","Data":"94d49d8097ae1fd76fdd4e47953421dfcf4ba9bd66f8dc14f0ffd65d153193df"} Dec 09 10:10:11 crc kubenswrapper[4824]: I1209 10:10:11.395622 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94d49d8097ae1fd76fdd4e47953421dfcf4ba9bd66f8dc14f0ffd65d153193df" Dec 09 10:10:11 crc kubenswrapper[4824]: I1209 10:10:11.398009 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"30eb761a-7bff-43a5-adfb-40e90e17e4a6","Type":"ContainerStarted","Data":"9cc814f255a3e8d2d480b4a92ba0edbcbf0c02ed24e819d5e91db3f835e461ff"} Dec 09 10:10:11 crc kubenswrapper[4824]: I1209 10:10:11.398206 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 09 10:10:11 crc kubenswrapper[4824]: I1209 10:10:11.401359 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"cee7a7d4-090b-4db0-abf1-d3f861f8e594","Type":"ContainerStarted","Data":"a32db2372ee433ebf981f7b113fb4bd4943f354f7d1ee3d51e9058cc39f25000"} Dec 09 10:10:11 crc kubenswrapper[4824]: I1209 10:10:11.401597 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-1" Dec 09 10:10:11 crc kubenswrapper[4824]: I1209 10:10:11.427002 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=15.386262203 podStartE2EDuration="1m50.426971463s" podCreationTimestamp="2025-12-09 10:08:21 +0000 UTC" firstStartedPulling="2025-12-09 10:08:35.50115473 +0000 UTC m=+1271.835659397" lastFinishedPulling="2025-12-09 10:10:10.54186399 +0000 UTC m=+1366.876368657" observedRunningTime="2025-12-09 10:10:11.421136727 +0000 UTC m=+1367.755641404" watchObservedRunningTime="2025-12-09 10:10:11.426971463 +0000 UTC m=+1367.761476130" Dec 09 10:10:11 crc kubenswrapper[4824]: I1209 10:10:11.447142 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 09 10:10:11 crc kubenswrapper[4824]: I1209 10:10:11.459053 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-1" podStartSLOduration=-9223371919.395746 podStartE2EDuration="1m57.459029224s" podCreationTimestamp="2025-12-09 10:08:14 +0000 UTC" firstStartedPulling="2025-12-09 10:08:17.137012552 +0000 UTC m=+1253.471517219" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:10:11.45798322 +0000 UTC m=+1367.792487917" watchObservedRunningTime="2025-12-09 10:10:11.459029224 +0000 UTC m=+1367.793533891" Dec 09 10:10:11 crc kubenswrapper[4824]: I1209 10:10:11.523284 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=-9223371919.33151 podStartE2EDuration="1m57.523265148s" podCreationTimestamp="2025-12-09 10:08:14 +0000 UTC" firstStartedPulling="2025-12-09 10:08:17.506155824 +0000 UTC m=+1253.840660491" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:10:11.51200899 +0000 UTC m=+1367.846513687" watchObservedRunningTime="2025-12-09 10:10:11.523265148 +0000 UTC m=+1367.857769815" Dec 09 10:10:12 crc kubenswrapper[4824]: I1209 10:10:12.413806 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f","Type":"ContainerStarted","Data":"fd451e0a617bb607e62796502aa6aa5bdbd756fed3cc947258bf5fbcfa6bb6c4"} Dec 09 10:10:13 crc kubenswrapper[4824]: I1209 10:10:13.223319 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:14 crc kubenswrapper[4824]: I1209 10:10:14.446760 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"8d47a992-6428-4729-8f4d-04af792e688e","Type":"ContainerStarted","Data":"2d99e9618d5d9984148ea2573701fb0fd09b8bb980843ed57ac98b1dd5b4c3e7"} Dec 09 10:10:14 crc kubenswrapper[4824]: I1209 10:10:14.503441 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f","Type":"ContainerStarted","Data":"b205a01e71b9f5a75a62f277d60913e10cc147fbf79f2af00be741ce06f905a3"} Dec 09 10:10:14 crc kubenswrapper[4824]: I1209 10:10:14.503489 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f","Type":"ContainerStarted","Data":"4c28f9331e200324ebd9ae0db55405a73192b636cf6f1aaa95ffc8a0663e351a"} Dec 09 10:10:14 crc kubenswrapper[4824]: I1209 10:10:14.503499 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f","Type":"ContainerStarted","Data":"2f491c8b434df69b75d884bd4b900c93175615219afdc9b875583c3a542ccf9d"} Dec 09 10:10:14 crc kubenswrapper[4824]: I1209 10:10:14.503508 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f","Type":"ContainerStarted","Data":"952d2252a6528e070be548359b629cea9699920edd8905b9fe4ef99e084e6b17"} Dec 09 10:10:14 crc kubenswrapper[4824]: I1209 10:10:14.529957 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=15.465373125 podStartE2EDuration="17.529929709s" podCreationTimestamp="2025-12-09 10:09:57 +0000 UTC" firstStartedPulling="2025-12-09 10:10:11.198060907 +0000 UTC m=+1367.532565574" lastFinishedPulling="2025-12-09 10:10:13.262617491 +0000 UTC m=+1369.597122158" observedRunningTime="2025-12-09 10:10:14.519741955 +0000 UTC m=+1370.854246622" watchObservedRunningTime="2025-12-09 10:10:14.529929709 +0000 UTC m=+1370.864434376" Dec 09 10:10:16 crc kubenswrapper[4824]: I1209 10:10:16.532388 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f","Type":"ContainerStarted","Data":"83cd0879550e3782e79b0c8326081e32a5c43cfcc2c74a292446ac38e8fa11d2"} Dec 09 10:10:16 crc kubenswrapper[4824]: I1209 10:10:16.533135 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f","Type":"ContainerStarted","Data":"d25fde79f0109e2c382c1b21f48d2772527013a985bea682c95203a8c3303cda"} Dec 09 10:10:16 crc kubenswrapper[4824]: I1209 10:10:16.533155 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f","Type":"ContainerStarted","Data":"782514abd32da3f9c399498c05e2f7ac1fa49c46bb0edaf899184b26434b4f29"} Dec 09 10:10:16 crc kubenswrapper[4824]: E1209 10:10:16.691709 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c1822c_1257_4d8c_9a17_d921929000fa.slice/crio-b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721e1031_d80c_406a_ad29_17879d58c3c3.slice/crio-conmon-88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c1822c_1257_4d8c_9a17_d921929000fa.slice/crio-conmon-b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721e1031_d80c_406a_ad29_17879d58c3c3.slice/crio-88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857.scope\": RecentStats: unable to find data in memory cache]" Dec 09 10:10:17 crc kubenswrapper[4824]: I1209 10:10:17.546181 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f","Type":"ContainerStarted","Data":"1cc90431cb3846a684358e6d946b1dd93ad6b01945da64be1a7dead3b8487418"} Dec 09 10:10:18 crc kubenswrapper[4824]: I1209 10:10:18.563459 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f","Type":"ContainerStarted","Data":"17a5647d7272f48a134af9ec0f689c54850df6bc3fac0deea1f4d78db4cc1b02"} Dec 09 10:10:18 crc kubenswrapper[4824]: I1209 10:10:18.564001 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f","Type":"ContainerStarted","Data":"5921782fc09946fb9ed15c21ff3fc718a2eafb8efb074d16168710c760b0417a"} Dec 09 10:10:18 crc kubenswrapper[4824]: I1209 10:10:18.564015 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f","Type":"ContainerStarted","Data":"91390ff4f60cb5e011bec19420f6af13cdc1e3e4dd1d29e39e97e889d7bcf770"} Dec 09 10:10:19 crc kubenswrapper[4824]: I1209 10:10:19.580230 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f","Type":"ContainerStarted","Data":"9d96cd0ee7bd30c04f2dc95e4e58241bdc93fa2890bb8972b1c9440121956b65"} Dec 09 10:10:19 crc kubenswrapper[4824]: I1209 10:10:19.580613 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f","Type":"ContainerStarted","Data":"1b05db108665b766c6dab821ccbe5f3f7f8989d3562910ed3bc518843ff799bc"} Dec 09 10:10:20 crc kubenswrapper[4824]: I1209 10:10:20.604152 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f","Type":"ContainerStarted","Data":"19bb231e78814f5cb2d763967ecf609ffb3b71443cedb4cb52891b0eb35bb834"} Dec 09 10:10:20 crc kubenswrapper[4824]: I1209 10:10:20.604624 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"66fd04f1-a0d9-4295-ac42-e5b20acc7c1f","Type":"ContainerStarted","Data":"465e16cc478c1c8d273a304bca340128162519d4692f617cbe2801d818e44319"} Dec 09 10:10:20 crc kubenswrapper[4824]: I1209 10:10:20.651726 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=53.379834722 podStartE2EDuration="59.651698781s" podCreationTimestamp="2025-12-09 10:09:21 +0000 UTC" firstStartedPulling="2025-12-09 10:10:11.459437157 +0000 UTC m=+1367.793941824" lastFinishedPulling="2025-12-09 10:10:17.731301226 +0000 UTC m=+1374.065805883" observedRunningTime="2025-12-09 10:10:20.647228809 +0000 UTC m=+1376.981733506" watchObservedRunningTime="2025-12-09 10:10:20.651698781 +0000 UTC m=+1376.986203448" Dec 09 10:10:20 crc kubenswrapper[4824]: I1209 10:10:20.930704 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-qbd6x"] Dec 09 10:10:20 crc kubenswrapper[4824]: E1209 10:10:20.931253 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abbc90b3-afb8-4c42-a07d-1d1c38b8d87c" containerName="swift-ring-rebalance" Dec 09 10:10:20 crc kubenswrapper[4824]: I1209 10:10:20.931276 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="abbc90b3-afb8-4c42-a07d-1d1c38b8d87c" containerName="swift-ring-rebalance" Dec 09 10:10:20 crc kubenswrapper[4824]: I1209 10:10:20.931534 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="abbc90b3-afb8-4c42-a07d-1d1c38b8d87c" containerName="swift-ring-rebalance" Dec 09 10:10:20 crc kubenswrapper[4824]: I1209 10:10:20.932875 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" Dec 09 10:10:20 crc kubenswrapper[4824]: I1209 10:10:20.935646 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 09 10:10:20 crc kubenswrapper[4824]: I1209 10:10:20.960453 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-qbd6x"] Dec 09 10:10:20 crc kubenswrapper[4824]: I1209 10:10:20.991986 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-qbd6x\" (UID: \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\") " pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" Dec 09 10:10:20 crc kubenswrapper[4824]: I1209 10:10:20.992039 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8drgj\" (UniqueName: \"kubernetes.io/projected/1ddb8431-cc07-48f4-a03f-7036462d0ba0-kube-api-access-8drgj\") pod \"dnsmasq-dns-764c5664d7-qbd6x\" (UID: \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\") " pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" Dec 09 10:10:20 crc kubenswrapper[4824]: I1209 10:10:20.992068 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-qbd6x\" (UID: \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\") " pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" Dec 09 10:10:20 crc kubenswrapper[4824]: I1209 10:10:20.992381 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-qbd6x\" (UID: \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\") " pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" Dec 09 10:10:20 crc kubenswrapper[4824]: I1209 10:10:20.992553 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-config\") pod \"dnsmasq-dns-764c5664d7-qbd6x\" (UID: \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\") " pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" Dec 09 10:10:20 crc kubenswrapper[4824]: I1209 10:10:20.992870 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-dns-svc\") pod \"dnsmasq-dns-764c5664d7-qbd6x\" (UID: \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\") " pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" Dec 09 10:10:21 crc kubenswrapper[4824]: I1209 10:10:21.094906 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-qbd6x\" (UID: \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\") " pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" Dec 09 10:10:21 crc kubenswrapper[4824]: I1209 10:10:21.094985 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8drgj\" (UniqueName: \"kubernetes.io/projected/1ddb8431-cc07-48f4-a03f-7036462d0ba0-kube-api-access-8drgj\") pod \"dnsmasq-dns-764c5664d7-qbd6x\" (UID: \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\") " pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" Dec 09 10:10:21 crc kubenswrapper[4824]: I1209 10:10:21.095016 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-qbd6x\" (UID: \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\") " pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" Dec 09 10:10:21 crc kubenswrapper[4824]: I1209 10:10:21.095092 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-qbd6x\" (UID: \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\") " pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" Dec 09 10:10:21 crc kubenswrapper[4824]: I1209 10:10:21.095131 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-config\") pod \"dnsmasq-dns-764c5664d7-qbd6x\" (UID: \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\") " pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" Dec 09 10:10:21 crc kubenswrapper[4824]: I1209 10:10:21.095184 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-dns-svc\") pod \"dnsmasq-dns-764c5664d7-qbd6x\" (UID: \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\") " pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" Dec 09 10:10:21 crc kubenswrapper[4824]: I1209 10:10:21.096084 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-dns-svc\") pod \"dnsmasq-dns-764c5664d7-qbd6x\" (UID: \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\") " pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" Dec 09 10:10:21 crc kubenswrapper[4824]: I1209 10:10:21.096564 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-qbd6x\" (UID: \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\") " pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" Dec 09 10:10:21 crc kubenswrapper[4824]: I1209 10:10:21.096812 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-qbd6x\" (UID: \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\") " pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" Dec 09 10:10:21 crc kubenswrapper[4824]: I1209 10:10:21.096830 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-qbd6x\" (UID: \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\") " pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" Dec 09 10:10:21 crc kubenswrapper[4824]: I1209 10:10:21.096853 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-config\") pod \"dnsmasq-dns-764c5664d7-qbd6x\" (UID: \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\") " pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" Dec 09 10:10:21 crc kubenswrapper[4824]: I1209 10:10:21.117338 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8drgj\" (UniqueName: \"kubernetes.io/projected/1ddb8431-cc07-48f4-a03f-7036462d0ba0-kube-api-access-8drgj\") pod \"dnsmasq-dns-764c5664d7-qbd6x\" (UID: \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\") " pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" Dec 09 10:10:21 crc kubenswrapper[4824]: I1209 10:10:21.254752 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" Dec 09 10:10:21 crc kubenswrapper[4824]: I1209 10:10:21.749450 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-qbd6x"] Dec 09 10:10:22 crc kubenswrapper[4824]: I1209 10:10:22.630035 4824 generic.go:334] "Generic (PLEG): container finished" podID="1ddb8431-cc07-48f4-a03f-7036462d0ba0" containerID="86b325e4f00388d45be553ffad5cfaf4ac0620cb023b426039f0dd08b64bc8a3" exitCode=0 Dec 09 10:10:22 crc kubenswrapper[4824]: I1209 10:10:22.630114 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" event={"ID":"1ddb8431-cc07-48f4-a03f-7036462d0ba0","Type":"ContainerDied","Data":"86b325e4f00388d45be553ffad5cfaf4ac0620cb023b426039f0dd08b64bc8a3"} Dec 09 10:10:22 crc kubenswrapper[4824]: I1209 10:10:22.630742 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" event={"ID":"1ddb8431-cc07-48f4-a03f-7036462d0ba0","Type":"ContainerStarted","Data":"8167486ab4d1393731aea993d1d6398986d33ce6c9aeea8203fd861d801e8827"} Dec 09 10:10:23 crc kubenswrapper[4824]: I1209 10:10:23.223740 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:23 crc kubenswrapper[4824]: I1209 10:10:23.228342 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:23 crc kubenswrapper[4824]: E1209 10:10:23.513030 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c1822c_1257_4d8c_9a17_d921929000fa.slice/crio-b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721e1031_d80c_406a_ad29_17879d58c3c3.slice/crio-88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c1822c_1257_4d8c_9a17_d921929000fa.slice/crio-conmon-b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod721e1031_d80c_406a_ad29_17879d58c3c3.slice/crio-conmon-88bc368ae81361a716ba4028a50297be7cab370f73db7984ebc8812995979857.scope\": RecentStats: unable to find data in memory cache]" Dec 09 10:10:23 crc kubenswrapper[4824]: I1209 10:10:23.647250 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" event={"ID":"1ddb8431-cc07-48f4-a03f-7036462d0ba0","Type":"ContainerStarted","Data":"1bbf83e9c9b5177e1c8227187951ee95a2c266a12029ff6996e1832cb7522afe"} Dec 09 10:10:23 crc kubenswrapper[4824]: I1209 10:10:23.650023 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-wbxqj" event={"ID":"80a3aebf-6b04-4e56-81cc-a632b53293b6","Type":"ContainerStarted","Data":"a5e2081631925849dc471ae1ff9d3dbd6d34197130fbfbbc666fc8fbd35e9fce"} Dec 09 10:10:23 crc kubenswrapper[4824]: I1209 10:10:23.651714 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:23 crc kubenswrapper[4824]: I1209 10:10:23.675265 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" podStartSLOduration=3.675236278 podStartE2EDuration="3.675236278s" podCreationTimestamp="2025-12-09 10:10:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:10:23.666431918 +0000 UTC m=+1380.000936605" watchObservedRunningTime="2025-12-09 10:10:23.675236278 +0000 UTC m=+1380.009740945" Dec 09 10:10:23 crc kubenswrapper[4824]: I1209 10:10:23.725556 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-wbxqj" podStartSLOduration=8.999282058 podStartE2EDuration="38.725533289s" podCreationTimestamp="2025-12-09 10:09:45 +0000 UTC" firstStartedPulling="2025-12-09 10:09:52.626588287 +0000 UTC m=+1348.961092954" lastFinishedPulling="2025-12-09 10:10:22.352839518 +0000 UTC m=+1378.687344185" observedRunningTime="2025-12-09 10:10:23.715708097 +0000 UTC m=+1380.050212774" watchObservedRunningTime="2025-12-09 10:10:23.725533289 +0000 UTC m=+1380.060037956" Dec 09 10:10:24 crc kubenswrapper[4824]: I1209 10:10:24.659546 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" Dec 09 10:10:25 crc kubenswrapper[4824]: I1209 10:10:25.841635 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-1" podUID="cee7a7d4-090b-4db0-abf1-d3f861f8e594" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.129:5671: connect: connection refused" Dec 09 10:10:26 crc kubenswrapper[4824]: I1209 10:10:26.188235 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="30eb761a-7bff-43a5-adfb-40e90e17e4a6" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.130:5671: connect: connection refused" Dec 09 10:10:26 crc kubenswrapper[4824]: I1209 10:10:26.688451 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 09 10:10:26 crc kubenswrapper[4824]: I1209 10:10:26.688838 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="8edb90f3-f84b-4700-993f-21ffa950b620" containerName="prometheus" containerID="cri-o://065fe900e020f72eec27757202af458fb43910dff8bc9cca1e8b163284ff096f" gracePeriod=600 Dec 09 10:10:26 crc kubenswrapper[4824]: I1209 10:10:26.688905 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="8edb90f3-f84b-4700-993f-21ffa950b620" containerName="thanos-sidecar" containerID="cri-o://99db94c0c3cfc2e1ecab46d0a7b0c3c3d0fcc1067806b3f5b0b09fba74941434" gracePeriod=600 Dec 09 10:10:26 crc kubenswrapper[4824]: I1209 10:10:26.688988 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="8edb90f3-f84b-4700-993f-21ffa950b620" containerName="config-reloader" containerID="cri-o://3bcac543b2d89d00cda20308eec14bbc4383f0f9d93d98669cd2beae0742c17f" gracePeriod=600 Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.693015 4824 generic.go:334] "Generic (PLEG): container finished" podID="8edb90f3-f84b-4700-993f-21ffa950b620" containerID="99db94c0c3cfc2e1ecab46d0a7b0c3c3d0fcc1067806b3f5b0b09fba74941434" exitCode=0 Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.694798 4824 generic.go:334] "Generic (PLEG): container finished" podID="8edb90f3-f84b-4700-993f-21ffa950b620" containerID="3bcac543b2d89d00cda20308eec14bbc4383f0f9d93d98669cd2beae0742c17f" exitCode=0 Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.694888 4824 generic.go:334] "Generic (PLEG): container finished" podID="8edb90f3-f84b-4700-993f-21ffa950b620" containerID="065fe900e020f72eec27757202af458fb43910dff8bc9cca1e8b163284ff096f" exitCode=0 Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.693094 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8edb90f3-f84b-4700-993f-21ffa950b620","Type":"ContainerDied","Data":"99db94c0c3cfc2e1ecab46d0a7b0c3c3d0fcc1067806b3f5b0b09fba74941434"} Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.695130 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8edb90f3-f84b-4700-993f-21ffa950b620","Type":"ContainerDied","Data":"3bcac543b2d89d00cda20308eec14bbc4383f0f9d93d98669cd2beae0742c17f"} Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.695212 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8edb90f3-f84b-4700-993f-21ffa950b620","Type":"ContainerDied","Data":"065fe900e020f72eec27757202af458fb43910dff8bc9cca1e8b163284ff096f"} Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.695293 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8edb90f3-f84b-4700-993f-21ffa950b620","Type":"ContainerDied","Data":"f3654a808c4c98871b6deba78bc55172067e7f26470defbce467d95bb738e7c3"} Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.695366 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3654a808c4c98871b6deba78bc55172067e7f26470defbce467d95bb738e7c3" Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.782738 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.944049 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8edb90f3-f84b-4700-993f-21ffa950b620-thanos-prometheus-http-client-file\") pod \"8edb90f3-f84b-4700-993f-21ffa950b620\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.944125 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8edb90f3-f84b-4700-993f-21ffa950b620-tls-assets\") pod \"8edb90f3-f84b-4700-993f-21ffa950b620\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.944163 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8edb90f3-f84b-4700-993f-21ffa950b620-prometheus-metric-storage-rulefiles-0\") pod \"8edb90f3-f84b-4700-993f-21ffa950b620\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.944249 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8edb90f3-f84b-4700-993f-21ffa950b620-config-out\") pod \"8edb90f3-f84b-4700-993f-21ffa950b620\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.944461 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a80b73-7299-419c-96ae-9bc835031b04\") pod \"8edb90f3-f84b-4700-993f-21ffa950b620\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.944595 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8edb90f3-f84b-4700-993f-21ffa950b620-config\") pod \"8edb90f3-f84b-4700-993f-21ffa950b620\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.945303 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-628jk\" (UniqueName: \"kubernetes.io/projected/8edb90f3-f84b-4700-993f-21ffa950b620-kube-api-access-628jk\") pod \"8edb90f3-f84b-4700-993f-21ffa950b620\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.945402 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8edb90f3-f84b-4700-993f-21ffa950b620-web-config\") pod \"8edb90f3-f84b-4700-993f-21ffa950b620\" (UID: \"8edb90f3-f84b-4700-993f-21ffa950b620\") " Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.945509 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8edb90f3-f84b-4700-993f-21ffa950b620-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "8edb90f3-f84b-4700-993f-21ffa950b620" (UID: "8edb90f3-f84b-4700-993f-21ffa950b620"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.946020 4824 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8edb90f3-f84b-4700-993f-21ffa950b620-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.955957 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8edb90f3-f84b-4700-993f-21ffa950b620-config-out" (OuterVolumeSpecName: "config-out") pod "8edb90f3-f84b-4700-993f-21ffa950b620" (UID: "8edb90f3-f84b-4700-993f-21ffa950b620"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.956232 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8edb90f3-f84b-4700-993f-21ffa950b620-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "8edb90f3-f84b-4700-993f-21ffa950b620" (UID: "8edb90f3-f84b-4700-993f-21ffa950b620"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.956593 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8edb90f3-f84b-4700-993f-21ffa950b620-config" (OuterVolumeSpecName: "config") pod "8edb90f3-f84b-4700-993f-21ffa950b620" (UID: "8edb90f3-f84b-4700-993f-21ffa950b620"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.991160 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8edb90f3-f84b-4700-993f-21ffa950b620-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "8edb90f3-f84b-4700-993f-21ffa950b620" (UID: "8edb90f3-f84b-4700-993f-21ffa950b620"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.997393 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8edb90f3-f84b-4700-993f-21ffa950b620-kube-api-access-628jk" (OuterVolumeSpecName: "kube-api-access-628jk") pod "8edb90f3-f84b-4700-993f-21ffa950b620" (UID: "8edb90f3-f84b-4700-993f-21ffa950b620"). InnerVolumeSpecName "kube-api-access-628jk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:10:27 crc kubenswrapper[4824]: I1209 10:10:27.998182 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a80b73-7299-419c-96ae-9bc835031b04" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "8edb90f3-f84b-4700-993f-21ffa950b620" (UID: "8edb90f3-f84b-4700-993f-21ffa950b620"). InnerVolumeSpecName "pvc-60a80b73-7299-419c-96ae-9bc835031b04". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.024422 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8edb90f3-f84b-4700-993f-21ffa950b620-web-config" (OuterVolumeSpecName: "web-config") pod "8edb90f3-f84b-4700-993f-21ffa950b620" (UID: "8edb90f3-f84b-4700-993f-21ffa950b620"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.055178 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8edb90f3-f84b-4700-993f-21ffa950b620-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.055255 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-628jk\" (UniqueName: \"kubernetes.io/projected/8edb90f3-f84b-4700-993f-21ffa950b620-kube-api-access-628jk\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.055274 4824 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8edb90f3-f84b-4700-993f-21ffa950b620-web-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.055289 4824 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8edb90f3-f84b-4700-993f-21ffa950b620-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.055309 4824 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8edb90f3-f84b-4700-993f-21ffa950b620-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.055322 4824 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8edb90f3-f84b-4700-993f-21ffa950b620-config-out\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.055367 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-60a80b73-7299-419c-96ae-9bc835031b04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a80b73-7299-419c-96ae-9bc835031b04\") on node \"crc\" " Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.105390 4824 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.105617 4824 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-60a80b73-7299-419c-96ae-9bc835031b04" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a80b73-7299-419c-96ae-9bc835031b04") on node "crc" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.158101 4824 reconciler_common.go:293] "Volume detached for volume \"pvc-60a80b73-7299-419c-96ae-9bc835031b04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a80b73-7299-419c-96ae-9bc835031b04\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.704322 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.750430 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.762600 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.781907 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 09 10:10:28 crc kubenswrapper[4824]: E1209 10:10:28.782688 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8edb90f3-f84b-4700-993f-21ffa950b620" containerName="thanos-sidecar" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.782767 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8edb90f3-f84b-4700-993f-21ffa950b620" containerName="thanos-sidecar" Dec 09 10:10:28 crc kubenswrapper[4824]: E1209 10:10:28.782867 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8edb90f3-f84b-4700-993f-21ffa950b620" containerName="prometheus" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.782925 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8edb90f3-f84b-4700-993f-21ffa950b620" containerName="prometheus" Dec 09 10:10:28 crc kubenswrapper[4824]: E1209 10:10:28.783010 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8edb90f3-f84b-4700-993f-21ffa950b620" containerName="config-reloader" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.783062 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8edb90f3-f84b-4700-993f-21ffa950b620" containerName="config-reloader" Dec 09 10:10:28 crc kubenswrapper[4824]: E1209 10:10:28.783127 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8edb90f3-f84b-4700-993f-21ffa950b620" containerName="init-config-reloader" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.783182 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8edb90f3-f84b-4700-993f-21ffa950b620" containerName="init-config-reloader" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.783460 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8edb90f3-f84b-4700-993f-21ffa950b620" containerName="config-reloader" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.783550 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8edb90f3-f84b-4700-993f-21ffa950b620" containerName="thanos-sidecar" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.783617 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8edb90f3-f84b-4700-993f-21ffa950b620" containerName="prometheus" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.785679 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.788659 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.789557 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.789676 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.790740 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.792585 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-qb6sm" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.793671 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.794621 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.794638 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.875382 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.875439 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.875475 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.875496 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdhmb\" (UniqueName: \"kubernetes.io/projected/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-kube-api-access-jdhmb\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.875587 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.875679 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.875712 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.875802 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.875830 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.875850 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-config\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.875883 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-60a80b73-7299-419c-96ae-9bc835031b04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a80b73-7299-419c-96ae-9bc835031b04\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.978131 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.978193 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdhmb\" (UniqueName: \"kubernetes.io/projected/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-kube-api-access-jdhmb\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.978248 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.978316 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.978354 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.978425 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.978450 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.978471 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-config\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.978513 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-60a80b73-7299-419c-96ae-9bc835031b04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a80b73-7299-419c-96ae-9bc835031b04\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.978614 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.978675 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.980108 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.981401 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.981450 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-60a80b73-7299-419c-96ae-9bc835031b04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a80b73-7299-419c-96ae-9bc835031b04\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8a9b576ec164efc98ef3fd54f8538cb024461023d7d2d915fde5b20b4af7c4e7/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.982883 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.983231 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.984169 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-config\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.984203 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.984515 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.984638 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.984800 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:28 crc kubenswrapper[4824]: I1209 10:10:28.991090 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:29 crc kubenswrapper[4824]: I1209 10:10:29.001366 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdhmb\" (UniqueName: \"kubernetes.io/projected/fb8f1e44-4c21-423a-bacc-1cca7d7715c3-kube-api-access-jdhmb\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:29 crc kubenswrapper[4824]: I1209 10:10:29.031015 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-60a80b73-7299-419c-96ae-9bc835031b04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a80b73-7299-419c-96ae-9bc835031b04\") pod \"prometheus-metric-storage-0\" (UID: \"fb8f1e44-4c21-423a-bacc-1cca7d7715c3\") " pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:29 crc kubenswrapper[4824]: I1209 10:10:29.109070 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:29 crc kubenswrapper[4824]: I1209 10:10:29.654633 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 09 10:10:29 crc kubenswrapper[4824]: I1209 10:10:29.716277 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"fb8f1e44-4c21-423a-bacc-1cca7d7715c3","Type":"ContainerStarted","Data":"9df9ffcd1de34f1f59b969f8f4b4683520e0c7f857827e998b40e9bad4469086"} Dec 09 10:10:29 crc kubenswrapper[4824]: I1209 10:10:29.924948 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8edb90f3-f84b-4700-993f-21ffa950b620" path="/var/lib/kubelet/pods/8edb90f3-f84b-4700-993f-21ffa950b620/volumes" Dec 09 10:10:31 crc kubenswrapper[4824]: I1209 10:10:31.256620 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" Dec 09 10:10:31 crc kubenswrapper[4824]: I1209 10:10:31.310873 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-t4gz7"] Dec 09 10:10:31 crc kubenswrapper[4824]: I1209 10:10:31.311135 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-t4gz7" podUID="82ce77fc-cdbd-4df3-b067-195040a2836e" containerName="dnsmasq-dns" containerID="cri-o://2b89f32a2604297e7699c708001e61817fa58b671867a19ac510149bc6732369" gracePeriod=10 Dec 09 10:10:31 crc kubenswrapper[4824]: I1209 10:10:31.741164 4824 generic.go:334] "Generic (PLEG): container finished" podID="82ce77fc-cdbd-4df3-b067-195040a2836e" containerID="2b89f32a2604297e7699c708001e61817fa58b671867a19ac510149bc6732369" exitCode=0 Dec 09 10:10:31 crc kubenswrapper[4824]: I1209 10:10:31.741516 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-t4gz7" event={"ID":"82ce77fc-cdbd-4df3-b067-195040a2836e","Type":"ContainerDied","Data":"2b89f32a2604297e7699c708001e61817fa58b671867a19ac510149bc6732369"} Dec 09 10:10:31 crc kubenswrapper[4824]: I1209 10:10:31.744571 4824 generic.go:334] "Generic (PLEG): container finished" podID="80a3aebf-6b04-4e56-81cc-a632b53293b6" containerID="a5e2081631925849dc471ae1ff9d3dbd6d34197130fbfbbc666fc8fbd35e9fce" exitCode=0 Dec 09 10:10:31 crc kubenswrapper[4824]: I1209 10:10:31.744619 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-wbxqj" event={"ID":"80a3aebf-6b04-4e56-81cc-a632b53293b6","Type":"ContainerDied","Data":"a5e2081631925849dc471ae1ff9d3dbd6d34197130fbfbbc666fc8fbd35e9fce"} Dec 09 10:10:32 crc kubenswrapper[4824]: I1209 10:10:32.045161 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-t4gz7" Dec 09 10:10:32 crc kubenswrapper[4824]: I1209 10:10:32.208598 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82ce77fc-cdbd-4df3-b067-195040a2836e-ovsdbserver-nb\") pod \"82ce77fc-cdbd-4df3-b067-195040a2836e\" (UID: \"82ce77fc-cdbd-4df3-b067-195040a2836e\") " Dec 09 10:10:32 crc kubenswrapper[4824]: I1209 10:10:32.209140 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82ce77fc-cdbd-4df3-b067-195040a2836e-ovsdbserver-sb\") pod \"82ce77fc-cdbd-4df3-b067-195040a2836e\" (UID: \"82ce77fc-cdbd-4df3-b067-195040a2836e\") " Dec 09 10:10:32 crc kubenswrapper[4824]: I1209 10:10:32.209184 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbnvd\" (UniqueName: \"kubernetes.io/projected/82ce77fc-cdbd-4df3-b067-195040a2836e-kube-api-access-rbnvd\") pod \"82ce77fc-cdbd-4df3-b067-195040a2836e\" (UID: \"82ce77fc-cdbd-4df3-b067-195040a2836e\") " Dec 09 10:10:32 crc kubenswrapper[4824]: I1209 10:10:32.209347 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82ce77fc-cdbd-4df3-b067-195040a2836e-dns-svc\") pod \"82ce77fc-cdbd-4df3-b067-195040a2836e\" (UID: \"82ce77fc-cdbd-4df3-b067-195040a2836e\") " Dec 09 10:10:32 crc kubenswrapper[4824]: I1209 10:10:32.209416 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82ce77fc-cdbd-4df3-b067-195040a2836e-config\") pod \"82ce77fc-cdbd-4df3-b067-195040a2836e\" (UID: \"82ce77fc-cdbd-4df3-b067-195040a2836e\") " Dec 09 10:10:32 crc kubenswrapper[4824]: I1209 10:10:32.216030 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82ce77fc-cdbd-4df3-b067-195040a2836e-kube-api-access-rbnvd" (OuterVolumeSpecName: "kube-api-access-rbnvd") pod "82ce77fc-cdbd-4df3-b067-195040a2836e" (UID: "82ce77fc-cdbd-4df3-b067-195040a2836e"). InnerVolumeSpecName "kube-api-access-rbnvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:10:32 crc kubenswrapper[4824]: I1209 10:10:32.263065 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82ce77fc-cdbd-4df3-b067-195040a2836e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "82ce77fc-cdbd-4df3-b067-195040a2836e" (UID: "82ce77fc-cdbd-4df3-b067-195040a2836e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:10:32 crc kubenswrapper[4824]: I1209 10:10:32.315690 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82ce77fc-cdbd-4df3-b067-195040a2836e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:32 crc kubenswrapper[4824]: I1209 10:10:32.315738 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbnvd\" (UniqueName: \"kubernetes.io/projected/82ce77fc-cdbd-4df3-b067-195040a2836e-kube-api-access-rbnvd\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:32 crc kubenswrapper[4824]: I1209 10:10:32.369460 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82ce77fc-cdbd-4df3-b067-195040a2836e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "82ce77fc-cdbd-4df3-b067-195040a2836e" (UID: "82ce77fc-cdbd-4df3-b067-195040a2836e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:10:32 crc kubenswrapper[4824]: I1209 10:10:32.417347 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82ce77fc-cdbd-4df3-b067-195040a2836e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:32 crc kubenswrapper[4824]: I1209 10:10:32.469206 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82ce77fc-cdbd-4df3-b067-195040a2836e-config" (OuterVolumeSpecName: "config") pod "82ce77fc-cdbd-4df3-b067-195040a2836e" (UID: "82ce77fc-cdbd-4df3-b067-195040a2836e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:10:32 crc kubenswrapper[4824]: I1209 10:10:32.471965 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82ce77fc-cdbd-4df3-b067-195040a2836e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "82ce77fc-cdbd-4df3-b067-195040a2836e" (UID: "82ce77fc-cdbd-4df3-b067-195040a2836e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:10:32 crc kubenswrapper[4824]: I1209 10:10:32.519535 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82ce77fc-cdbd-4df3-b067-195040a2836e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:32 crc kubenswrapper[4824]: I1209 10:10:32.519607 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82ce77fc-cdbd-4df3-b067-195040a2836e-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:32 crc kubenswrapper[4824]: I1209 10:10:32.757238 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-t4gz7" event={"ID":"82ce77fc-cdbd-4df3-b067-195040a2836e","Type":"ContainerDied","Data":"11ad6ca9d90a564448cdb01df3ff58d259af1eb8991df735eea3faabbe2714da"} Dec 09 10:10:32 crc kubenswrapper[4824]: I1209 10:10:32.757285 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-t4gz7" Dec 09 10:10:32 crc kubenswrapper[4824]: I1209 10:10:32.757330 4824 scope.go:117] "RemoveContainer" containerID="2b89f32a2604297e7699c708001e61817fa58b671867a19ac510149bc6732369" Dec 09 10:10:32 crc kubenswrapper[4824]: I1209 10:10:32.786574 4824 scope.go:117] "RemoveContainer" containerID="0123721178eca4cbfc6b2f6749af4123c08e9b695c4bd8681e0c0fc9b3126303" Dec 09 10:10:32 crc kubenswrapper[4824]: I1209 10:10:32.797507 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-t4gz7"] Dec 09 10:10:32 crc kubenswrapper[4824]: I1209 10:10:32.818159 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-t4gz7"] Dec 09 10:10:33 crc kubenswrapper[4824]: I1209 10:10:33.378116 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-wbxqj" Dec 09 10:10:33 crc kubenswrapper[4824]: I1209 10:10:33.449934 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80a3aebf-6b04-4e56-81cc-a632b53293b6-config-data\") pod \"80a3aebf-6b04-4e56-81cc-a632b53293b6\" (UID: \"80a3aebf-6b04-4e56-81cc-a632b53293b6\") " Dec 09 10:10:33 crc kubenswrapper[4824]: I1209 10:10:33.450003 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80a3aebf-6b04-4e56-81cc-a632b53293b6-combined-ca-bundle\") pod \"80a3aebf-6b04-4e56-81cc-a632b53293b6\" (UID: \"80a3aebf-6b04-4e56-81cc-a632b53293b6\") " Dec 09 10:10:33 crc kubenswrapper[4824]: I1209 10:10:33.450084 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/80a3aebf-6b04-4e56-81cc-a632b53293b6-db-sync-config-data\") pod \"80a3aebf-6b04-4e56-81cc-a632b53293b6\" (UID: \"80a3aebf-6b04-4e56-81cc-a632b53293b6\") " Dec 09 10:10:33 crc kubenswrapper[4824]: I1209 10:10:33.457731 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80a3aebf-6b04-4e56-81cc-a632b53293b6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "80a3aebf-6b04-4e56-81cc-a632b53293b6" (UID: "80a3aebf-6b04-4e56-81cc-a632b53293b6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:10:33 crc kubenswrapper[4824]: I1209 10:10:33.488111 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80a3aebf-6b04-4e56-81cc-a632b53293b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "80a3aebf-6b04-4e56-81cc-a632b53293b6" (UID: "80a3aebf-6b04-4e56-81cc-a632b53293b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:10:33 crc kubenswrapper[4824]: I1209 10:10:33.509574 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80a3aebf-6b04-4e56-81cc-a632b53293b6-config-data" (OuterVolumeSpecName: "config-data") pod "80a3aebf-6b04-4e56-81cc-a632b53293b6" (UID: "80a3aebf-6b04-4e56-81cc-a632b53293b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:10:33 crc kubenswrapper[4824]: I1209 10:10:33.552080 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6gxs\" (UniqueName: \"kubernetes.io/projected/80a3aebf-6b04-4e56-81cc-a632b53293b6-kube-api-access-p6gxs\") pod \"80a3aebf-6b04-4e56-81cc-a632b53293b6\" (UID: \"80a3aebf-6b04-4e56-81cc-a632b53293b6\") " Dec 09 10:10:33 crc kubenswrapper[4824]: I1209 10:10:33.552400 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80a3aebf-6b04-4e56-81cc-a632b53293b6-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:33 crc kubenswrapper[4824]: I1209 10:10:33.552422 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80a3aebf-6b04-4e56-81cc-a632b53293b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:33 crc kubenswrapper[4824]: I1209 10:10:33.552435 4824 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/80a3aebf-6b04-4e56-81cc-a632b53293b6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:33 crc kubenswrapper[4824]: I1209 10:10:33.555551 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80a3aebf-6b04-4e56-81cc-a632b53293b6-kube-api-access-p6gxs" (OuterVolumeSpecName: "kube-api-access-p6gxs") pod "80a3aebf-6b04-4e56-81cc-a632b53293b6" (UID: "80a3aebf-6b04-4e56-81cc-a632b53293b6"). InnerVolumeSpecName "kube-api-access-p6gxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:10:33 crc kubenswrapper[4824]: I1209 10:10:33.654904 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6gxs\" (UniqueName: \"kubernetes.io/projected/80a3aebf-6b04-4e56-81cc-a632b53293b6-kube-api-access-p6gxs\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:33 crc kubenswrapper[4824]: I1209 10:10:33.803483 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-wbxqj" Dec 09 10:10:33 crc kubenswrapper[4824]: I1209 10:10:33.803485 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-wbxqj" event={"ID":"80a3aebf-6b04-4e56-81cc-a632b53293b6","Type":"ContainerDied","Data":"d9f1f3112821b81ab615cb06293cd5943dbedb7fe1be424fe7a4c30c50c7c103"} Dec 09 10:10:33 crc kubenswrapper[4824]: I1209 10:10:33.804369 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9f1f3112821b81ab615cb06293cd5943dbedb7fe1be424fe7a4c30c50c7c103" Dec 09 10:10:33 crc kubenswrapper[4824]: I1209 10:10:33.813456 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"fb8f1e44-4c21-423a-bacc-1cca7d7715c3","Type":"ContainerStarted","Data":"a047a4100142e8f715dd650ab74ced9ec383da240c7aa970068b418abeb5d3a9"} Dec 09 10:10:33 crc kubenswrapper[4824]: I1209 10:10:33.924028 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82ce77fc-cdbd-4df3-b067-195040a2836e" path="/var/lib/kubelet/pods/82ce77fc-cdbd-4df3-b067-195040a2836e/volumes" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.185454 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-fgj5w"] Dec 09 10:10:34 crc kubenswrapper[4824]: E1209 10:10:34.188097 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82ce77fc-cdbd-4df3-b067-195040a2836e" containerName="init" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.188345 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="82ce77fc-cdbd-4df3-b067-195040a2836e" containerName="init" Dec 09 10:10:34 crc kubenswrapper[4824]: E1209 10:10:34.188454 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82ce77fc-cdbd-4df3-b067-195040a2836e" containerName="dnsmasq-dns" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.188526 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="82ce77fc-cdbd-4df3-b067-195040a2836e" containerName="dnsmasq-dns" Dec 09 10:10:34 crc kubenswrapper[4824]: E1209 10:10:34.188632 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80a3aebf-6b04-4e56-81cc-a632b53293b6" containerName="glance-db-sync" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.188712 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="80a3aebf-6b04-4e56-81cc-a632b53293b6" containerName="glance-db-sync" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.189173 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="82ce77fc-cdbd-4df3-b067-195040a2836e" containerName="dnsmasq-dns" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.189265 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="80a3aebf-6b04-4e56-81cc-a632b53293b6" containerName="glance-db-sync" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.190917 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.198477 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-fgj5w"] Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.373404 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-fgj5w\" (UID: \"6a45d593-c64c-471e-b0fa-171446548a4d\") " pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.373519 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-fgj5w\" (UID: \"6a45d593-c64c-471e-b0fa-171446548a4d\") " pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.373557 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-fgj5w\" (UID: \"6a45d593-c64c-471e-b0fa-171446548a4d\") " pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.373597 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-fgj5w\" (UID: \"6a45d593-c64c-471e-b0fa-171446548a4d\") " pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.374094 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q44x9\" (UniqueName: \"kubernetes.io/projected/6a45d593-c64c-471e-b0fa-171446548a4d-kube-api-access-q44x9\") pod \"dnsmasq-dns-74f6bcbc87-fgj5w\" (UID: \"6a45d593-c64c-471e-b0fa-171446548a4d\") " pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.374310 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-config\") pod \"dnsmasq-dns-74f6bcbc87-fgj5w\" (UID: \"6a45d593-c64c-471e-b0fa-171446548a4d\") " pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.475983 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q44x9\" (UniqueName: \"kubernetes.io/projected/6a45d593-c64c-471e-b0fa-171446548a4d-kube-api-access-q44x9\") pod \"dnsmasq-dns-74f6bcbc87-fgj5w\" (UID: \"6a45d593-c64c-471e-b0fa-171446548a4d\") " pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.476051 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-config\") pod \"dnsmasq-dns-74f6bcbc87-fgj5w\" (UID: \"6a45d593-c64c-471e-b0fa-171446548a4d\") " pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.476108 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-fgj5w\" (UID: \"6a45d593-c64c-471e-b0fa-171446548a4d\") " pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.476164 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-fgj5w\" (UID: \"6a45d593-c64c-471e-b0fa-171446548a4d\") " pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.476188 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-fgj5w\" (UID: \"6a45d593-c64c-471e-b0fa-171446548a4d\") " pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.476208 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-fgj5w\" (UID: \"6a45d593-c64c-471e-b0fa-171446548a4d\") " pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.477110 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-config\") pod \"dnsmasq-dns-74f6bcbc87-fgj5w\" (UID: \"6a45d593-c64c-471e-b0fa-171446548a4d\") " pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.477254 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-fgj5w\" (UID: \"6a45d593-c64c-471e-b0fa-171446548a4d\") " pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.477554 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-fgj5w\" (UID: \"6a45d593-c64c-471e-b0fa-171446548a4d\") " pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.477861 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-fgj5w\" (UID: \"6a45d593-c64c-471e-b0fa-171446548a4d\") " pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.477983 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-fgj5w\" (UID: \"6a45d593-c64c-471e-b0fa-171446548a4d\") " pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.500438 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q44x9\" (UniqueName: \"kubernetes.io/projected/6a45d593-c64c-471e-b0fa-171446548a4d-kube-api-access-q44x9\") pod \"dnsmasq-dns-74f6bcbc87-fgj5w\" (UID: \"6a45d593-c64c-471e-b0fa-171446548a4d\") " pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:10:34 crc kubenswrapper[4824]: I1209 10:10:34.518039 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:10:35 crc kubenswrapper[4824]: I1209 10:10:35.121670 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-fgj5w"] Dec 09 10:10:35 crc kubenswrapper[4824]: W1209 10:10:35.143947 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a45d593_c64c_471e_b0fa_171446548a4d.slice/crio-182446cf60766ff1ce3f29478da994c779b28b093149949b54447f66562c814f WatchSource:0}: Error finding container 182446cf60766ff1ce3f29478da994c779b28b093149949b54447f66562c814f: Status 404 returned error can't find the container with id 182446cf60766ff1ce3f29478da994c779b28b093149949b54447f66562c814f Dec 09 10:10:35 crc kubenswrapper[4824]: I1209 10:10:35.841045 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-1" Dec 09 10:10:35 crc kubenswrapper[4824]: I1209 10:10:35.845549 4824 generic.go:334] "Generic (PLEG): container finished" podID="6a45d593-c64c-471e-b0fa-171446548a4d" containerID="0de4dfe5dc112545d3c8f9f2f646791151f090c0f7071a68110442d9a180004b" exitCode=0 Dec 09 10:10:35 crc kubenswrapper[4824]: I1209 10:10:35.845612 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" event={"ID":"6a45d593-c64c-471e-b0fa-171446548a4d","Type":"ContainerDied","Data":"0de4dfe5dc112545d3c8f9f2f646791151f090c0f7071a68110442d9a180004b"} Dec 09 10:10:35 crc kubenswrapper[4824]: I1209 10:10:35.845678 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" event={"ID":"6a45d593-c64c-471e-b0fa-171446548a4d","Type":"ContainerStarted","Data":"182446cf60766ff1ce3f29478da994c779b28b093149949b54447f66562c814f"} Dec 09 10:10:36 crc kubenswrapper[4824]: I1209 10:10:36.189375 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 09 10:10:36 crc kubenswrapper[4824]: I1209 10:10:36.901550 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" event={"ID":"6a45d593-c64c-471e-b0fa-171446548a4d","Type":"ContainerStarted","Data":"085feec377c96e6eba63a801f66855826edb5591113a6501dd0576d7f8ea05d0"} Dec 09 10:10:36 crc kubenswrapper[4824]: I1209 10:10:36.903549 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:10:36 crc kubenswrapper[4824]: I1209 10:10:36.933517 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" podStartSLOduration=2.933495171 podStartE2EDuration="2.933495171s" podCreationTimestamp="2025-12-09 10:10:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:10:36.925898899 +0000 UTC m=+1393.260403596" watchObservedRunningTime="2025-12-09 10:10:36.933495171 +0000 UTC m=+1393.267999838" Dec 09 10:10:38 crc kubenswrapper[4824]: I1209 10:10:38.923042 4824 generic.go:334] "Generic (PLEG): container finished" podID="fb8f1e44-4c21-423a-bacc-1cca7d7715c3" containerID="a047a4100142e8f715dd650ab74ced9ec383da240c7aa970068b418abeb5d3a9" exitCode=0 Dec 09 10:10:38 crc kubenswrapper[4824]: I1209 10:10:38.923140 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"fb8f1e44-4c21-423a-bacc-1cca7d7715c3","Type":"ContainerDied","Data":"a047a4100142e8f715dd650ab74ced9ec383da240c7aa970068b418abeb5d3a9"} Dec 09 10:10:39 crc kubenswrapper[4824]: I1209 10:10:39.940088 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"fb8f1e44-4c21-423a-bacc-1cca7d7715c3","Type":"ContainerStarted","Data":"adae8d93e4c50f4e5637886919ec9af1d2e54909a2d926a833550cf678e0007d"} Dec 09 10:10:42 crc kubenswrapper[4824]: I1209 10:10:42.989935 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"fb8f1e44-4c21-423a-bacc-1cca7d7715c3","Type":"ContainerStarted","Data":"60ead2a86262ced45a54c4e6bcf50545489e18b0b4b97f7ef1fa96db5b967b79"} Dec 09 10:10:42 crc kubenswrapper[4824]: I1209 10:10:42.990609 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"fb8f1e44-4c21-423a-bacc-1cca7d7715c3","Type":"ContainerStarted","Data":"de23f9078df0537fb1c5ed6c0ae561171f3e871f652ff5ce83b8bebe3121ef8b"} Dec 09 10:10:43 crc kubenswrapper[4824]: I1209 10:10:43.026436 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=15.026403255 podStartE2EDuration="15.026403255s" podCreationTimestamp="2025-12-09 10:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:10:43.01648781 +0000 UTC m=+1399.350992497" watchObservedRunningTime="2025-12-09 10:10:43.026403255 +0000 UTC m=+1399.360907922" Dec 09 10:10:44 crc kubenswrapper[4824]: I1209 10:10:44.109852 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:44 crc kubenswrapper[4824]: I1209 10:10:44.110261 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:44 crc kubenswrapper[4824]: I1209 10:10:44.116033 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:44 crc kubenswrapper[4824]: I1209 10:10:44.520043 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:10:44 crc kubenswrapper[4824]: I1209 10:10:44.641162 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-qbd6x"] Dec 09 10:10:44 crc kubenswrapper[4824]: I1209 10:10:44.641872 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" podUID="1ddb8431-cc07-48f4-a03f-7036462d0ba0" containerName="dnsmasq-dns" containerID="cri-o://1bbf83e9c9b5177e1c8227187951ee95a2c266a12029ff6996e1832cb7522afe" gracePeriod=10 Dec 09 10:10:45 crc kubenswrapper[4824]: I1209 10:10:45.016007 4824 generic.go:334] "Generic (PLEG): container finished" podID="09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" containerID="17519c5dfbefe63576e185d8b281c4f8a3c5c8eaff5cd9a05b5db8801ca6f9b2" exitCode=0 Dec 09 10:10:45 crc kubenswrapper[4824]: I1209 10:10:45.016080 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55","Type":"ContainerDied","Data":"17519c5dfbefe63576e185d8b281c4f8a3c5c8eaff5cd9a05b5db8801ca6f9b2"} Dec 09 10:10:45 crc kubenswrapper[4824]: I1209 10:10:45.018543 4824 generic.go:334] "Generic (PLEG): container finished" podID="1ddb8431-cc07-48f4-a03f-7036462d0ba0" containerID="1bbf83e9c9b5177e1c8227187951ee95a2c266a12029ff6996e1832cb7522afe" exitCode=0 Dec 09 10:10:45 crc kubenswrapper[4824]: I1209 10:10:45.019720 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" event={"ID":"1ddb8431-cc07-48f4-a03f-7036462d0ba0","Type":"ContainerDied","Data":"1bbf83e9c9b5177e1c8227187951ee95a2c266a12029ff6996e1832cb7522afe"} Dec 09 10:10:45 crc kubenswrapper[4824]: I1209 10:10:45.030142 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 09 10:10:45 crc kubenswrapper[4824]: I1209 10:10:45.385641 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" Dec 09 10:10:45 crc kubenswrapper[4824]: I1209 10:10:45.498490 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-config\") pod \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\" (UID: \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\") " Dec 09 10:10:45 crc kubenswrapper[4824]: I1209 10:10:45.500120 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-dns-swift-storage-0\") pod \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\" (UID: \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\") " Dec 09 10:10:45 crc kubenswrapper[4824]: I1209 10:10:45.500213 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-ovsdbserver-nb\") pod \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\" (UID: \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\") " Dec 09 10:10:45 crc kubenswrapper[4824]: I1209 10:10:45.500299 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8drgj\" (UniqueName: \"kubernetes.io/projected/1ddb8431-cc07-48f4-a03f-7036462d0ba0-kube-api-access-8drgj\") pod \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\" (UID: \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\") " Dec 09 10:10:45 crc kubenswrapper[4824]: I1209 10:10:45.500447 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-dns-svc\") pod \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\" (UID: \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\") " Dec 09 10:10:45 crc kubenswrapper[4824]: I1209 10:10:45.500574 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-ovsdbserver-sb\") pod \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\" (UID: \"1ddb8431-cc07-48f4-a03f-7036462d0ba0\") " Dec 09 10:10:45 crc kubenswrapper[4824]: I1209 10:10:45.520227 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ddb8431-cc07-48f4-a03f-7036462d0ba0-kube-api-access-8drgj" (OuterVolumeSpecName: "kube-api-access-8drgj") pod "1ddb8431-cc07-48f4-a03f-7036462d0ba0" (UID: "1ddb8431-cc07-48f4-a03f-7036462d0ba0"). InnerVolumeSpecName "kube-api-access-8drgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:10:45 crc kubenswrapper[4824]: I1209 10:10:45.572740 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1ddb8431-cc07-48f4-a03f-7036462d0ba0" (UID: "1ddb8431-cc07-48f4-a03f-7036462d0ba0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:10:45 crc kubenswrapper[4824]: I1209 10:10:45.594647 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1ddb8431-cc07-48f4-a03f-7036462d0ba0" (UID: "1ddb8431-cc07-48f4-a03f-7036462d0ba0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:10:45 crc kubenswrapper[4824]: I1209 10:10:45.598593 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1ddb8431-cc07-48f4-a03f-7036462d0ba0" (UID: "1ddb8431-cc07-48f4-a03f-7036462d0ba0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:10:45 crc kubenswrapper[4824]: I1209 10:10:45.605595 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:45 crc kubenswrapper[4824]: I1209 10:10:45.605706 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8drgj\" (UniqueName: \"kubernetes.io/projected/1ddb8431-cc07-48f4-a03f-7036462d0ba0-kube-api-access-8drgj\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:45 crc kubenswrapper[4824]: I1209 10:10:45.605730 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:45 crc kubenswrapper[4824]: I1209 10:10:45.605750 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:45 crc kubenswrapper[4824]: I1209 10:10:45.614549 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1ddb8431-cc07-48f4-a03f-7036462d0ba0" (UID: "1ddb8431-cc07-48f4-a03f-7036462d0ba0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:10:45 crc kubenswrapper[4824]: I1209 10:10:45.619761 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-config" (OuterVolumeSpecName: "config") pod "1ddb8431-cc07-48f4-a03f-7036462d0ba0" (UID: "1ddb8431-cc07-48f4-a03f-7036462d0ba0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:10:45 crc kubenswrapper[4824]: I1209 10:10:45.708871 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:45 crc kubenswrapper[4824]: I1209 10:10:45.708919 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ddb8431-cc07-48f4-a03f-7036462d0ba0-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:10:46 crc kubenswrapper[4824]: I1209 10:10:46.031024 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55","Type":"ContainerStarted","Data":"3ad239061116629b299b08f30e913aae04ed870beb2645bdf1ab9cfc59480f62"} Dec 09 10:10:46 crc kubenswrapper[4824]: I1209 10:10:46.032533 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-2" Dec 09 10:10:46 crc kubenswrapper[4824]: I1209 10:10:46.033889 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" event={"ID":"1ddb8431-cc07-48f4-a03f-7036462d0ba0","Type":"ContainerDied","Data":"8167486ab4d1393731aea993d1d6398986d33ce6c9aeea8203fd861d801e8827"} Dec 09 10:10:46 crc kubenswrapper[4824]: I1209 10:10:46.033945 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-qbd6x" Dec 09 10:10:46 crc kubenswrapper[4824]: I1209 10:10:46.033951 4824 scope.go:117] "RemoveContainer" containerID="1bbf83e9c9b5177e1c8227187951ee95a2c266a12029ff6996e1832cb7522afe" Dec 09 10:10:46 crc kubenswrapper[4824]: I1209 10:10:46.068558 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-2" podStartSLOduration=-9223371884.786242 podStartE2EDuration="2m32.068533404s" podCreationTimestamp="2025-12-09 10:08:14 +0000 UTC" firstStartedPulling="2025-12-09 10:08:16.702992484 +0000 UTC m=+1253.037497151" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:10:46.067083178 +0000 UTC m=+1402.401587855" watchObservedRunningTime="2025-12-09 10:10:46.068533404 +0000 UTC m=+1402.403038071" Dec 09 10:10:46 crc kubenswrapper[4824]: I1209 10:10:46.088146 4824 scope.go:117] "RemoveContainer" containerID="86b325e4f00388d45be553ffad5cfaf4ac0620cb023b426039f0dd08b64bc8a3" Dec 09 10:10:46 crc kubenswrapper[4824]: I1209 10:10:46.106915 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-qbd6x"] Dec 09 10:10:46 crc kubenswrapper[4824]: I1209 10:10:46.136603 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-qbd6x"] Dec 09 10:10:47 crc kubenswrapper[4824]: I1209 10:10:47.929001 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ddb8431-cc07-48f4-a03f-7036462d0ba0" path="/var/lib/kubelet/pods/1ddb8431-cc07-48f4-a03f-7036462d0ba0/volumes" Dec 09 10:10:55 crc kubenswrapper[4824]: I1209 10:10:55.931833 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-2" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.493692 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-zh7x5"] Dec 09 10:10:56 crc kubenswrapper[4824]: E1209 10:10:56.495112 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ddb8431-cc07-48f4-a03f-7036462d0ba0" containerName="dnsmasq-dns" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.495131 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ddb8431-cc07-48f4-a03f-7036462d0ba0" containerName="dnsmasq-dns" Dec 09 10:10:56 crc kubenswrapper[4824]: E1209 10:10:56.495198 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ddb8431-cc07-48f4-a03f-7036462d0ba0" containerName="init" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.495204 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ddb8431-cc07-48f4-a03f-7036462d0ba0" containerName="init" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.495622 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ddb8431-cc07-48f4-a03f-7036462d0ba0" containerName="dnsmasq-dns" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.497272 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-zh7x5" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.520238 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-zh7x5"] Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.676411 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-nxcqz"] Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.678284 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nxcqz" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.693318 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mppnk\" (UniqueName: \"kubernetes.io/projected/55e299a6-a633-4e32-9766-54cbff8053b0-kube-api-access-mppnk\") pod \"heat-db-create-zh7x5\" (UID: \"55e299a6-a633-4e32-9766-54cbff8053b0\") " pod="openstack/heat-db-create-zh7x5" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.693739 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55e299a6-a633-4e32-9766-54cbff8053b0-operator-scripts\") pod \"heat-db-create-zh7x5\" (UID: \"55e299a6-a633-4e32-9766-54cbff8053b0\") " pod="openstack/heat-db-create-zh7x5" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.695176 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-nxcqz"] Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.767894 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-9mmtj"] Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.769954 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-9mmtj" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.777137 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-719f-account-create-update-mlf5c"] Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.778816 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-719f-account-create-update-mlf5c" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.786166 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-9mmtj"] Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.795489 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55e299a6-a633-4e32-9766-54cbff8053b0-operator-scripts\") pod \"heat-db-create-zh7x5\" (UID: \"55e299a6-a633-4e32-9766-54cbff8053b0\") " pod="openstack/heat-db-create-zh7x5" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.795619 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ee07f30-f716-4d7a-b5dd-71de0e128d3f-operator-scripts\") pod \"barbican-db-create-nxcqz\" (UID: \"7ee07f30-f716-4d7a-b5dd-71de0e128d3f\") " pod="openstack/barbican-db-create-nxcqz" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.795681 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-719f-account-create-update-mlf5c"] Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.795728 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mppnk\" (UniqueName: \"kubernetes.io/projected/55e299a6-a633-4e32-9766-54cbff8053b0-kube-api-access-mppnk\") pod \"heat-db-create-zh7x5\" (UID: \"55e299a6-a633-4e32-9766-54cbff8053b0\") " pod="openstack/heat-db-create-zh7x5" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.795933 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldtsb\" (UniqueName: \"kubernetes.io/projected/7ee07f30-f716-4d7a-b5dd-71de0e128d3f-kube-api-access-ldtsb\") pod \"barbican-db-create-nxcqz\" (UID: \"7ee07f30-f716-4d7a-b5dd-71de0e128d3f\") " pod="openstack/barbican-db-create-nxcqz" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.796892 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.797185 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55e299a6-a633-4e32-9766-54cbff8053b0-operator-scripts\") pod \"heat-db-create-zh7x5\" (UID: \"55e299a6-a633-4e32-9766-54cbff8053b0\") " pod="openstack/heat-db-create-zh7x5" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.853812 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mppnk\" (UniqueName: \"kubernetes.io/projected/55e299a6-a633-4e32-9766-54cbff8053b0-kube-api-access-mppnk\") pod \"heat-db-create-zh7x5\" (UID: \"55e299a6-a633-4e32-9766-54cbff8053b0\") " pod="openstack/heat-db-create-zh7x5" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.902828 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ee07f30-f716-4d7a-b5dd-71de0e128d3f-operator-scripts\") pod \"barbican-db-create-nxcqz\" (UID: \"7ee07f30-f716-4d7a-b5dd-71de0e128d3f\") " pod="openstack/barbican-db-create-nxcqz" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.902909 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2npvb\" (UniqueName: \"kubernetes.io/projected/48b5dcdc-81a1-49f9-b5b6-5c9445be07dc-kube-api-access-2npvb\") pod \"heat-719f-account-create-update-mlf5c\" (UID: \"48b5dcdc-81a1-49f9-b5b6-5c9445be07dc\") " pod="openstack/heat-719f-account-create-update-mlf5c" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.902956 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79dm8\" (UniqueName: \"kubernetes.io/projected/ae040f9f-f187-4477-9455-af94eda40a1a-kube-api-access-79dm8\") pod \"cinder-db-create-9mmtj\" (UID: \"ae040f9f-f187-4477-9455-af94eda40a1a\") " pod="openstack/cinder-db-create-9mmtj" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.903088 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldtsb\" (UniqueName: \"kubernetes.io/projected/7ee07f30-f716-4d7a-b5dd-71de0e128d3f-kube-api-access-ldtsb\") pod \"barbican-db-create-nxcqz\" (UID: \"7ee07f30-f716-4d7a-b5dd-71de0e128d3f\") " pod="openstack/barbican-db-create-nxcqz" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.903179 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48b5dcdc-81a1-49f9-b5b6-5c9445be07dc-operator-scripts\") pod \"heat-719f-account-create-update-mlf5c\" (UID: \"48b5dcdc-81a1-49f9-b5b6-5c9445be07dc\") " pod="openstack/heat-719f-account-create-update-mlf5c" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.903207 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae040f9f-f187-4477-9455-af94eda40a1a-operator-scripts\") pod \"cinder-db-create-9mmtj\" (UID: \"ae040f9f-f187-4477-9455-af94eda40a1a\") " pod="openstack/cinder-db-create-9mmtj" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.904213 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ee07f30-f716-4d7a-b5dd-71de0e128d3f-operator-scripts\") pod \"barbican-db-create-nxcqz\" (UID: \"7ee07f30-f716-4d7a-b5dd-71de0e128d3f\") " pod="openstack/barbican-db-create-nxcqz" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.904872 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-3340-account-create-update-fskfw"] Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.906593 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3340-account-create-update-fskfw" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.918213 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.931615 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-3340-account-create-update-fskfw"] Dec 09 10:10:56 crc kubenswrapper[4824]: I1209 10:10:56.938601 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-zh7x5" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:56.942464 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldtsb\" (UniqueName: \"kubernetes.io/projected/7ee07f30-f716-4d7a-b5dd-71de0e128d3f-kube-api-access-ldtsb\") pod \"barbican-db-create-nxcqz\" (UID: \"7ee07f30-f716-4d7a-b5dd-71de0e128d3f\") " pod="openstack/barbican-db-create-nxcqz" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.205270 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nxcqz" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.211238 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48b5dcdc-81a1-49f9-b5b6-5c9445be07dc-operator-scripts\") pod \"heat-719f-account-create-update-mlf5c\" (UID: \"48b5dcdc-81a1-49f9-b5b6-5c9445be07dc\") " pod="openstack/heat-719f-account-create-update-mlf5c" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.211319 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae040f9f-f187-4477-9455-af94eda40a1a-operator-scripts\") pod \"cinder-db-create-9mmtj\" (UID: \"ae040f9f-f187-4477-9455-af94eda40a1a\") " pod="openstack/cinder-db-create-9mmtj" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.211428 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2npvb\" (UniqueName: \"kubernetes.io/projected/48b5dcdc-81a1-49f9-b5b6-5c9445be07dc-kube-api-access-2npvb\") pod \"heat-719f-account-create-update-mlf5c\" (UID: \"48b5dcdc-81a1-49f9-b5b6-5c9445be07dc\") " pod="openstack/heat-719f-account-create-update-mlf5c" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.211462 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79dm8\" (UniqueName: \"kubernetes.io/projected/ae040f9f-f187-4477-9455-af94eda40a1a-kube-api-access-79dm8\") pod \"cinder-db-create-9mmtj\" (UID: \"ae040f9f-f187-4477-9455-af94eda40a1a\") " pod="openstack/cinder-db-create-9mmtj" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.212919 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48b5dcdc-81a1-49f9-b5b6-5c9445be07dc-operator-scripts\") pod \"heat-719f-account-create-update-mlf5c\" (UID: \"48b5dcdc-81a1-49f9-b5b6-5c9445be07dc\") " pod="openstack/heat-719f-account-create-update-mlf5c" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.213470 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae040f9f-f187-4477-9455-af94eda40a1a-operator-scripts\") pod \"cinder-db-create-9mmtj\" (UID: \"ae040f9f-f187-4477-9455-af94eda40a1a\") " pod="openstack/cinder-db-create-9mmtj" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.237341 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2npvb\" (UniqueName: \"kubernetes.io/projected/48b5dcdc-81a1-49f9-b5b6-5c9445be07dc-kube-api-access-2npvb\") pod \"heat-719f-account-create-update-mlf5c\" (UID: \"48b5dcdc-81a1-49f9-b5b6-5c9445be07dc\") " pod="openstack/heat-719f-account-create-update-mlf5c" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.249714 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79dm8\" (UniqueName: \"kubernetes.io/projected/ae040f9f-f187-4477-9455-af94eda40a1a-kube-api-access-79dm8\") pod \"cinder-db-create-9mmtj\" (UID: \"ae040f9f-f187-4477-9455-af94eda40a1a\") " pod="openstack/cinder-db-create-9mmtj" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.314148 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a2cda53-48d6-480c-9da8-1caec867aee5-operator-scripts\") pod \"cinder-3340-account-create-update-fskfw\" (UID: \"8a2cda53-48d6-480c-9da8-1caec867aee5\") " pod="openstack/cinder-3340-account-create-update-fskfw" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.314609 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6tlp\" (UniqueName: \"kubernetes.io/projected/8a2cda53-48d6-480c-9da8-1caec867aee5-kube-api-access-d6tlp\") pod \"cinder-3340-account-create-update-fskfw\" (UID: \"8a2cda53-48d6-480c-9da8-1caec867aee5\") " pod="openstack/cinder-3340-account-create-update-fskfw" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.328056 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-6swf9"] Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.329742 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-6swf9" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.337438 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.343457 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.343717 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ggbf4" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.343904 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.355200 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-6swf9"] Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.382358 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-0196-account-create-update-llvmc"] Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.384214 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0196-account-create-update-llvmc" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.388458 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.399349 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-9mmtj" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.453062 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a2cda53-48d6-480c-9da8-1caec867aee5-operator-scripts\") pod \"cinder-3340-account-create-update-fskfw\" (UID: \"8a2cda53-48d6-480c-9da8-1caec867aee5\") " pod="openstack/cinder-3340-account-create-update-fskfw" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.453152 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6tlp\" (UniqueName: \"kubernetes.io/projected/8a2cda53-48d6-480c-9da8-1caec867aee5-kube-api-access-d6tlp\") pod \"cinder-3340-account-create-update-fskfw\" (UID: \"8a2cda53-48d6-480c-9da8-1caec867aee5\") " pod="openstack/cinder-3340-account-create-update-fskfw" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.454209 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a2cda53-48d6-480c-9da8-1caec867aee5-operator-scripts\") pod \"cinder-3340-account-create-update-fskfw\" (UID: \"8a2cda53-48d6-480c-9da8-1caec867aee5\") " pod="openstack/cinder-3340-account-create-update-fskfw" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.466294 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-9px5r"] Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.467992 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9px5r" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.480347 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6tlp\" (UniqueName: \"kubernetes.io/projected/8a2cda53-48d6-480c-9da8-1caec867aee5-kube-api-access-d6tlp\") pod \"cinder-3340-account-create-update-fskfw\" (UID: \"8a2cda53-48d6-480c-9da8-1caec867aee5\") " pod="openstack/cinder-3340-account-create-update-fskfw" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.483131 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-719f-account-create-update-mlf5c" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.517828 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-9px5r"] Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.543284 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-0196-account-create-update-llvmc"] Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.543992 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3340-account-create-update-fskfw" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.558666 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/826695ab-42b3-496a-a67b-add8ab95d4c1-operator-scripts\") pod \"barbican-0196-account-create-update-llvmc\" (UID: \"826695ab-42b3-496a-a67b-add8ab95d4c1\") " pod="openstack/barbican-0196-account-create-update-llvmc" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.558998 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da292326-c56e-42d0-bfe2-2b75b53c210d-config-data\") pod \"keystone-db-sync-6swf9\" (UID: \"da292326-c56e-42d0-bfe2-2b75b53c210d\") " pod="openstack/keystone-db-sync-6swf9" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.559026 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w82g\" (UniqueName: \"kubernetes.io/projected/da292326-c56e-42d0-bfe2-2b75b53c210d-kube-api-access-7w82g\") pod \"keystone-db-sync-6swf9\" (UID: \"da292326-c56e-42d0-bfe2-2b75b53c210d\") " pod="openstack/keystone-db-sync-6swf9" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.559121 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2vp2\" (UniqueName: \"kubernetes.io/projected/826695ab-42b3-496a-a67b-add8ab95d4c1-kube-api-access-p2vp2\") pod \"barbican-0196-account-create-update-llvmc\" (UID: \"826695ab-42b3-496a-a67b-add8ab95d4c1\") " pod="openstack/barbican-0196-account-create-update-llvmc" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.559188 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da292326-c56e-42d0-bfe2-2b75b53c210d-combined-ca-bundle\") pod \"keystone-db-sync-6swf9\" (UID: \"da292326-c56e-42d0-bfe2-2b75b53c210d\") " pod="openstack/keystone-db-sync-6swf9" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.663900 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/826695ab-42b3-496a-a67b-add8ab95d4c1-operator-scripts\") pod \"barbican-0196-account-create-update-llvmc\" (UID: \"826695ab-42b3-496a-a67b-add8ab95d4c1\") " pod="openstack/barbican-0196-account-create-update-llvmc" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.663967 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da292326-c56e-42d0-bfe2-2b75b53c210d-config-data\") pod \"keystone-db-sync-6swf9\" (UID: \"da292326-c56e-42d0-bfe2-2b75b53c210d\") " pod="openstack/keystone-db-sync-6swf9" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.664004 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bcqg\" (UniqueName: \"kubernetes.io/projected/f6a3773f-0af2-4536-a361-dd51cedd700f-kube-api-access-9bcqg\") pod \"neutron-db-create-9px5r\" (UID: \"f6a3773f-0af2-4536-a361-dd51cedd700f\") " pod="openstack/neutron-db-create-9px5r" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.664036 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w82g\" (UniqueName: \"kubernetes.io/projected/da292326-c56e-42d0-bfe2-2b75b53c210d-kube-api-access-7w82g\") pod \"keystone-db-sync-6swf9\" (UID: \"da292326-c56e-42d0-bfe2-2b75b53c210d\") " pod="openstack/keystone-db-sync-6swf9" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.664157 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2vp2\" (UniqueName: \"kubernetes.io/projected/826695ab-42b3-496a-a67b-add8ab95d4c1-kube-api-access-p2vp2\") pod \"barbican-0196-account-create-update-llvmc\" (UID: \"826695ab-42b3-496a-a67b-add8ab95d4c1\") " pod="openstack/barbican-0196-account-create-update-llvmc" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.664260 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da292326-c56e-42d0-bfe2-2b75b53c210d-combined-ca-bundle\") pod \"keystone-db-sync-6swf9\" (UID: \"da292326-c56e-42d0-bfe2-2b75b53c210d\") " pod="openstack/keystone-db-sync-6swf9" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.664328 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6a3773f-0af2-4536-a361-dd51cedd700f-operator-scripts\") pod \"neutron-db-create-9px5r\" (UID: \"f6a3773f-0af2-4536-a361-dd51cedd700f\") " pod="openstack/neutron-db-create-9px5r" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.665902 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/826695ab-42b3-496a-a67b-add8ab95d4c1-operator-scripts\") pod \"barbican-0196-account-create-update-llvmc\" (UID: \"826695ab-42b3-496a-a67b-add8ab95d4c1\") " pod="openstack/barbican-0196-account-create-update-llvmc" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.683854 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da292326-c56e-42d0-bfe2-2b75b53c210d-combined-ca-bundle\") pod \"keystone-db-sync-6swf9\" (UID: \"da292326-c56e-42d0-bfe2-2b75b53c210d\") " pod="openstack/keystone-db-sync-6swf9" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.684024 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da292326-c56e-42d0-bfe2-2b75b53c210d-config-data\") pod \"keystone-db-sync-6swf9\" (UID: \"da292326-c56e-42d0-bfe2-2b75b53c210d\") " pod="openstack/keystone-db-sync-6swf9" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.690713 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w82g\" (UniqueName: \"kubernetes.io/projected/da292326-c56e-42d0-bfe2-2b75b53c210d-kube-api-access-7w82g\") pod \"keystone-db-sync-6swf9\" (UID: \"da292326-c56e-42d0-bfe2-2b75b53c210d\") " pod="openstack/keystone-db-sync-6swf9" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.693236 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-b08a-account-create-update-jx676"] Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.695603 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b08a-account-create-update-jx676" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.697549 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2vp2\" (UniqueName: \"kubernetes.io/projected/826695ab-42b3-496a-a67b-add8ab95d4c1-kube-api-access-p2vp2\") pod \"barbican-0196-account-create-update-llvmc\" (UID: \"826695ab-42b3-496a-a67b-add8ab95d4c1\") " pod="openstack/barbican-0196-account-create-update-llvmc" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.699910 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.708091 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b08a-account-create-update-jx676"] Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.774296 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6a3773f-0af2-4536-a361-dd51cedd700f-operator-scripts\") pod \"neutron-db-create-9px5r\" (UID: \"f6a3773f-0af2-4536-a361-dd51cedd700f\") " pod="openstack/neutron-db-create-9px5r" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.774551 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bcqg\" (UniqueName: \"kubernetes.io/projected/f6a3773f-0af2-4536-a361-dd51cedd700f-kube-api-access-9bcqg\") pod \"neutron-db-create-9px5r\" (UID: \"f6a3773f-0af2-4536-a361-dd51cedd700f\") " pod="openstack/neutron-db-create-9px5r" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.776308 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6a3773f-0af2-4536-a361-dd51cedd700f-operator-scripts\") pod \"neutron-db-create-9px5r\" (UID: \"f6a3773f-0af2-4536-a361-dd51cedd700f\") " pod="openstack/neutron-db-create-9px5r" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.808021 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-6swf9" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.823526 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0196-account-create-update-llvmc" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.832647 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bcqg\" (UniqueName: \"kubernetes.io/projected/f6a3773f-0af2-4536-a361-dd51cedd700f-kube-api-access-9bcqg\") pod \"neutron-db-create-9px5r\" (UID: \"f6a3773f-0af2-4536-a361-dd51cedd700f\") " pod="openstack/neutron-db-create-9px5r" Dec 09 10:10:57 crc kubenswrapper[4824]: I1209 10:10:57.844045 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9px5r" Dec 09 10:10:58 crc kubenswrapper[4824]: I1209 10:10:58.123425 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55a79bdb-6f53-4f93-85cd-e674210cc7e7-operator-scripts\") pod \"neutron-b08a-account-create-update-jx676\" (UID: \"55a79bdb-6f53-4f93-85cd-e674210cc7e7\") " pod="openstack/neutron-b08a-account-create-update-jx676" Dec 09 10:10:58 crc kubenswrapper[4824]: I1209 10:10:58.123518 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5fmv\" (UniqueName: \"kubernetes.io/projected/55a79bdb-6f53-4f93-85cd-e674210cc7e7-kube-api-access-b5fmv\") pod \"neutron-b08a-account-create-update-jx676\" (UID: \"55a79bdb-6f53-4f93-85cd-e674210cc7e7\") " pod="openstack/neutron-b08a-account-create-update-jx676" Dec 09 10:10:58 crc kubenswrapper[4824]: I1209 10:10:58.227520 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5fmv\" (UniqueName: \"kubernetes.io/projected/55a79bdb-6f53-4f93-85cd-e674210cc7e7-kube-api-access-b5fmv\") pod \"neutron-b08a-account-create-update-jx676\" (UID: \"55a79bdb-6f53-4f93-85cd-e674210cc7e7\") " pod="openstack/neutron-b08a-account-create-update-jx676" Dec 09 10:10:58 crc kubenswrapper[4824]: I1209 10:10:58.227759 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55a79bdb-6f53-4f93-85cd-e674210cc7e7-operator-scripts\") pod \"neutron-b08a-account-create-update-jx676\" (UID: \"55a79bdb-6f53-4f93-85cd-e674210cc7e7\") " pod="openstack/neutron-b08a-account-create-update-jx676" Dec 09 10:10:58 crc kubenswrapper[4824]: I1209 10:10:58.229146 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55a79bdb-6f53-4f93-85cd-e674210cc7e7-operator-scripts\") pod \"neutron-b08a-account-create-update-jx676\" (UID: \"55a79bdb-6f53-4f93-85cd-e674210cc7e7\") " pod="openstack/neutron-b08a-account-create-update-jx676" Dec 09 10:10:58 crc kubenswrapper[4824]: I1209 10:10:58.302095 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5fmv\" (UniqueName: \"kubernetes.io/projected/55a79bdb-6f53-4f93-85cd-e674210cc7e7-kube-api-access-b5fmv\") pod \"neutron-b08a-account-create-update-jx676\" (UID: \"55a79bdb-6f53-4f93-85cd-e674210cc7e7\") " pod="openstack/neutron-b08a-account-create-update-jx676" Dec 09 10:10:58 crc kubenswrapper[4824]: I1209 10:10:58.408620 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b08a-account-create-update-jx676" Dec 09 10:10:58 crc kubenswrapper[4824]: I1209 10:10:58.664582 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-zh7x5"] Dec 09 10:10:58 crc kubenswrapper[4824]: I1209 10:10:58.715837 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-nxcqz"] Dec 09 10:10:58 crc kubenswrapper[4824]: W1209 10:10:58.830902 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55e299a6_a633_4e32_9766_54cbff8053b0.slice/crio-ca4f8974e87ebd586b31f40786a15ddbd93cf215e4425617c4a5342c681451ef WatchSource:0}: Error finding container ca4f8974e87ebd586b31f40786a15ddbd93cf215e4425617c4a5342c681451ef: Status 404 returned error can't find the container with id ca4f8974e87ebd586b31f40786a15ddbd93cf215e4425617c4a5342c681451ef Dec 09 10:10:58 crc kubenswrapper[4824]: W1209 10:10:58.894172 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ee07f30_f716_4d7a_b5dd_71de0e128d3f.slice/crio-b50356ce8581ade212cd45137fd2c8c55a17a5b1c9a842507228a672e16eff9a WatchSource:0}: Error finding container b50356ce8581ade212cd45137fd2c8c55a17a5b1c9a842507228a672e16eff9a: Status 404 returned error can't find the container with id b50356ce8581ade212cd45137fd2c8c55a17a5b1c9a842507228a672e16eff9a Dec 09 10:10:59 crc kubenswrapper[4824]: I1209 10:10:59.586517 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nxcqz" event={"ID":"7ee07f30-f716-4d7a-b5dd-71de0e128d3f","Type":"ContainerStarted","Data":"b50356ce8581ade212cd45137fd2c8c55a17a5b1c9a842507228a672e16eff9a"} Dec 09 10:10:59 crc kubenswrapper[4824]: I1209 10:10:59.604376 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-zh7x5" event={"ID":"55e299a6-a633-4e32-9766-54cbff8053b0","Type":"ContainerStarted","Data":"ca4f8974e87ebd586b31f40786a15ddbd93cf215e4425617c4a5342c681451ef"} Dec 09 10:10:59 crc kubenswrapper[4824]: I1209 10:10:59.621117 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-create-zh7x5" podStartSLOduration=3.621057744 podStartE2EDuration="3.621057744s" podCreationTimestamp="2025-12-09 10:10:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:10:59.620241417 +0000 UTC m=+1415.954746084" watchObservedRunningTime="2025-12-09 10:10:59.621057744 +0000 UTC m=+1415.955562411" Dec 09 10:10:59 crc kubenswrapper[4824]: I1209 10:10:59.694807 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-3340-account-create-update-fskfw"] Dec 09 10:11:00 crc kubenswrapper[4824]: W1209 10:11:00.014089 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae040f9f_f187_4477_9455_af94eda40a1a.slice/crio-6f1be4155832bc397848c30ec15e85fddfb2236173335068afc14bf85853c0f5 WatchSource:0}: Error finding container 6f1be4155832bc397848c30ec15e85fddfb2236173335068afc14bf85853c0f5: Status 404 returned error can't find the container with id 6f1be4155832bc397848c30ec15e85fddfb2236173335068afc14bf85853c0f5 Dec 09 10:11:00 crc kubenswrapper[4824]: I1209 10:11:00.067403 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-9mmtj"] Dec 09 10:11:00 crc kubenswrapper[4824]: I1209 10:11:00.074332 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-719f-account-create-update-mlf5c"] Dec 09 10:11:00 crc kubenswrapper[4824]: I1209 10:11:00.103639 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-9px5r"] Dec 09 10:11:00 crc kubenswrapper[4824]: W1209 10:11:00.124004 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6a3773f_0af2_4536_a361_dd51cedd700f.slice/crio-aff45440dcdeb5478ac01bebab61b3cd4b87b1ce6ff500a4014306a6508f4c8c WatchSource:0}: Error finding container aff45440dcdeb5478ac01bebab61b3cd4b87b1ce6ff500a4014306a6508f4c8c: Status 404 returned error can't find the container with id aff45440dcdeb5478ac01bebab61b3cd4b87b1ce6ff500a4014306a6508f4c8c Dec 09 10:11:00 crc kubenswrapper[4824]: I1209 10:11:00.179518 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-6swf9"] Dec 09 10:11:00 crc kubenswrapper[4824]: W1209 10:11:00.238918 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda292326_c56e_42d0_bfe2_2b75b53c210d.slice/crio-70ac9f5b948a7c677d0c3ad55d67442affab350383f6b2aa57006ae9932329d5 WatchSource:0}: Error finding container 70ac9f5b948a7c677d0c3ad55d67442affab350383f6b2aa57006ae9932329d5: Status 404 returned error can't find the container with id 70ac9f5b948a7c677d0c3ad55d67442affab350383f6b2aa57006ae9932329d5 Dec 09 10:11:00 crc kubenswrapper[4824]: I1209 10:11:00.440358 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b08a-account-create-update-jx676"] Dec 09 10:11:00 crc kubenswrapper[4824]: W1209 10:11:00.456352 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55a79bdb_6f53_4f93_85cd_e674210cc7e7.slice/crio-7b26c83199062e59c7c027ef919d3888780d2e59210d952ea41bfe241f5717ff WatchSource:0}: Error finding container 7b26c83199062e59c7c027ef919d3888780d2e59210d952ea41bfe241f5717ff: Status 404 returned error can't find the container with id 7b26c83199062e59c7c027ef919d3888780d2e59210d952ea41bfe241f5717ff Dec 09 10:11:00 crc kubenswrapper[4824]: I1209 10:11:00.526996 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-0196-account-create-update-llvmc"] Dec 09 10:11:00 crc kubenswrapper[4824]: I1209 10:11:00.632516 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9px5r" event={"ID":"f6a3773f-0af2-4536-a361-dd51cedd700f","Type":"ContainerStarted","Data":"aff45440dcdeb5478ac01bebab61b3cd4b87b1ce6ff500a4014306a6508f4c8c"} Dec 09 10:11:00 crc kubenswrapper[4824]: I1209 10:11:00.638274 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b08a-account-create-update-jx676" event={"ID":"55a79bdb-6f53-4f93-85cd-e674210cc7e7","Type":"ContainerStarted","Data":"7b26c83199062e59c7c027ef919d3888780d2e59210d952ea41bfe241f5717ff"} Dec 09 10:11:00 crc kubenswrapper[4824]: I1209 10:11:00.645461 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-6swf9" event={"ID":"da292326-c56e-42d0-bfe2-2b75b53c210d","Type":"ContainerStarted","Data":"70ac9f5b948a7c677d0c3ad55d67442affab350383f6b2aa57006ae9932329d5"} Dec 09 10:11:00 crc kubenswrapper[4824]: I1209 10:11:00.652282 4824 generic.go:334] "Generic (PLEG): container finished" podID="55e299a6-a633-4e32-9766-54cbff8053b0" containerID="d056fadb93420fa399bb35c0d9edfbda3533432b04327f7d6c526b8f3f20f64c" exitCode=0 Dec 09 10:11:00 crc kubenswrapper[4824]: I1209 10:11:00.652415 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-zh7x5" event={"ID":"55e299a6-a633-4e32-9766-54cbff8053b0","Type":"ContainerDied","Data":"d056fadb93420fa399bb35c0d9edfbda3533432b04327f7d6c526b8f3f20f64c"} Dec 09 10:11:00 crc kubenswrapper[4824]: I1209 10:11:00.657442 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-9mmtj" event={"ID":"ae040f9f-f187-4477-9455-af94eda40a1a","Type":"ContainerStarted","Data":"6f1be4155832bc397848c30ec15e85fddfb2236173335068afc14bf85853c0f5"} Dec 09 10:11:00 crc kubenswrapper[4824]: I1209 10:11:00.661287 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-719f-account-create-update-mlf5c" event={"ID":"48b5dcdc-81a1-49f9-b5b6-5c9445be07dc","Type":"ContainerStarted","Data":"d42df4baf36fb684468cd910ef6fabd3db82bc040a61d91edb8840eafc9f1ce1"} Dec 09 10:11:00 crc kubenswrapper[4824]: I1209 10:11:00.664730 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3340-account-create-update-fskfw" event={"ID":"8a2cda53-48d6-480c-9da8-1caec867aee5","Type":"ContainerStarted","Data":"dafa8bf8cf3a340f7ac03ec0b5a7990b5fa9c89716729d5c7b9820fab179799e"} Dec 09 10:11:00 crc kubenswrapper[4824]: I1209 10:11:00.667339 4824 generic.go:334] "Generic (PLEG): container finished" podID="7ee07f30-f716-4d7a-b5dd-71de0e128d3f" containerID="f886bcd18faabc96f773b58bea74aa161fc7d6fa281648ab17854ca99895392f" exitCode=0 Dec 09 10:11:00 crc kubenswrapper[4824]: I1209 10:11:00.667426 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nxcqz" event={"ID":"7ee07f30-f716-4d7a-b5dd-71de0e128d3f","Type":"ContainerDied","Data":"f886bcd18faabc96f773b58bea74aa161fc7d6fa281648ab17854ca99895392f"} Dec 09 10:11:00 crc kubenswrapper[4824]: I1209 10:11:00.694919 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-9mmtj" podStartSLOduration=4.694892673 podStartE2EDuration="4.694892673s" podCreationTimestamp="2025-12-09 10:10:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:11:00.686504946 +0000 UTC m=+1417.021009633" watchObservedRunningTime="2025-12-09 10:11:00.694892673 +0000 UTC m=+1417.029397340" Dec 09 10:11:01 crc kubenswrapper[4824]: I1209 10:11:01.692706 4824 generic.go:334] "Generic (PLEG): container finished" podID="8a2cda53-48d6-480c-9da8-1caec867aee5" containerID="6acbb8552ca41469eedb970660ce6259b53fc161ce19ea9368d90e4325235c02" exitCode=0 Dec 09 10:11:01 crc kubenswrapper[4824]: I1209 10:11:01.692865 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3340-account-create-update-fskfw" event={"ID":"8a2cda53-48d6-480c-9da8-1caec867aee5","Type":"ContainerDied","Data":"6acbb8552ca41469eedb970660ce6259b53fc161ce19ea9368d90e4325235c02"} Dec 09 10:11:01 crc kubenswrapper[4824]: I1209 10:11:01.694797 4824 generic.go:334] "Generic (PLEG): container finished" podID="f6a3773f-0af2-4536-a361-dd51cedd700f" containerID="445d9b2d240dbce5259e65a1dd08ca31ec4fe19d06a1709bf06377d6f083ebde" exitCode=0 Dec 09 10:11:01 crc kubenswrapper[4824]: I1209 10:11:01.694851 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9px5r" event={"ID":"f6a3773f-0af2-4536-a361-dd51cedd700f","Type":"ContainerDied","Data":"445d9b2d240dbce5259e65a1dd08ca31ec4fe19d06a1709bf06377d6f083ebde"} Dec 09 10:11:01 crc kubenswrapper[4824]: I1209 10:11:01.698160 4824 generic.go:334] "Generic (PLEG): container finished" podID="826695ab-42b3-496a-a67b-add8ab95d4c1" containerID="2d511d51e8516b43cd422d16759aa763e3a7165dc038bbd12bcb6f11d8cccb8f" exitCode=0 Dec 09 10:11:01 crc kubenswrapper[4824]: I1209 10:11:01.698315 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0196-account-create-update-llvmc" event={"ID":"826695ab-42b3-496a-a67b-add8ab95d4c1","Type":"ContainerDied","Data":"2d511d51e8516b43cd422d16759aa763e3a7165dc038bbd12bcb6f11d8cccb8f"} Dec 09 10:11:01 crc kubenswrapper[4824]: I1209 10:11:01.698394 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0196-account-create-update-llvmc" event={"ID":"826695ab-42b3-496a-a67b-add8ab95d4c1","Type":"ContainerStarted","Data":"d52c8e2c2279b8e82884a2f9c7911d242003cfb8fd1fd30a45f0ff04593bff5a"} Dec 09 10:11:01 crc kubenswrapper[4824]: I1209 10:11:01.701003 4824 generic.go:334] "Generic (PLEG): container finished" podID="55a79bdb-6f53-4f93-85cd-e674210cc7e7" containerID="3685b228d8569d658b8bb0a9ef88c43b270f4ad1d48e5b2abcacf4d6a2c84a3f" exitCode=0 Dec 09 10:11:01 crc kubenswrapper[4824]: I1209 10:11:01.701148 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b08a-account-create-update-jx676" event={"ID":"55a79bdb-6f53-4f93-85cd-e674210cc7e7","Type":"ContainerDied","Data":"3685b228d8569d658b8bb0a9ef88c43b270f4ad1d48e5b2abcacf4d6a2c84a3f"} Dec 09 10:11:01 crc kubenswrapper[4824]: I1209 10:11:01.704117 4824 generic.go:334] "Generic (PLEG): container finished" podID="ae040f9f-f187-4477-9455-af94eda40a1a" containerID="6d8a450a8b8d26e3ab8e5ca1f2315522285cd246059130be9689bb6ecf8274a7" exitCode=0 Dec 09 10:11:01 crc kubenswrapper[4824]: I1209 10:11:01.704157 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-9mmtj" event={"ID":"ae040f9f-f187-4477-9455-af94eda40a1a","Type":"ContainerDied","Data":"6d8a450a8b8d26e3ab8e5ca1f2315522285cd246059130be9689bb6ecf8274a7"} Dec 09 10:11:01 crc kubenswrapper[4824]: I1209 10:11:01.706565 4824 generic.go:334] "Generic (PLEG): container finished" podID="48b5dcdc-81a1-49f9-b5b6-5c9445be07dc" containerID="a8d7ef59965802baa03a3321640fe5629f7313e2ac4298b53df52384cae5caa9" exitCode=0 Dec 09 10:11:01 crc kubenswrapper[4824]: I1209 10:11:01.706896 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-719f-account-create-update-mlf5c" event={"ID":"48b5dcdc-81a1-49f9-b5b6-5c9445be07dc","Type":"ContainerDied","Data":"a8d7ef59965802baa03a3321640fe5629f7313e2ac4298b53df52384cae5caa9"} Dec 09 10:11:02 crc kubenswrapper[4824]: I1209 10:11:02.466584 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nxcqz" Dec 09 10:11:02 crc kubenswrapper[4824]: I1209 10:11:02.560766 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ee07f30-f716-4d7a-b5dd-71de0e128d3f-operator-scripts\") pod \"7ee07f30-f716-4d7a-b5dd-71de0e128d3f\" (UID: \"7ee07f30-f716-4d7a-b5dd-71de0e128d3f\") " Dec 09 10:11:02 crc kubenswrapper[4824]: I1209 10:11:02.560923 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldtsb\" (UniqueName: \"kubernetes.io/projected/7ee07f30-f716-4d7a-b5dd-71de0e128d3f-kube-api-access-ldtsb\") pod \"7ee07f30-f716-4d7a-b5dd-71de0e128d3f\" (UID: \"7ee07f30-f716-4d7a-b5dd-71de0e128d3f\") " Dec 09 10:11:02 crc kubenswrapper[4824]: I1209 10:11:02.562290 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ee07f30-f716-4d7a-b5dd-71de0e128d3f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7ee07f30-f716-4d7a-b5dd-71de0e128d3f" (UID: "7ee07f30-f716-4d7a-b5dd-71de0e128d3f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:11:02 crc kubenswrapper[4824]: I1209 10:11:02.570745 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ee07f30-f716-4d7a-b5dd-71de0e128d3f-kube-api-access-ldtsb" (OuterVolumeSpecName: "kube-api-access-ldtsb") pod "7ee07f30-f716-4d7a-b5dd-71de0e128d3f" (UID: "7ee07f30-f716-4d7a-b5dd-71de0e128d3f"). InnerVolumeSpecName "kube-api-access-ldtsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:11:02 crc kubenswrapper[4824]: I1209 10:11:02.663509 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ee07f30-f716-4d7a-b5dd-71de0e128d3f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:02 crc kubenswrapper[4824]: I1209 10:11:02.663565 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldtsb\" (UniqueName: \"kubernetes.io/projected/7ee07f30-f716-4d7a-b5dd-71de0e128d3f-kube-api-access-ldtsb\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:02 crc kubenswrapper[4824]: I1209 10:11:02.679912 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-zh7x5" Dec 09 10:11:03 crc kubenswrapper[4824]: I1209 10:11:03.055364 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:11:03 crc kubenswrapper[4824]: I1209 10:11:03.055961 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:11:03 crc kubenswrapper[4824]: I1209 10:11:03.091180 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55e299a6-a633-4e32-9766-54cbff8053b0-operator-scripts\") pod \"55e299a6-a633-4e32-9766-54cbff8053b0\" (UID: \"55e299a6-a633-4e32-9766-54cbff8053b0\") " Dec 09 10:11:03 crc kubenswrapper[4824]: I1209 10:11:03.091665 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mppnk\" (UniqueName: \"kubernetes.io/projected/55e299a6-a633-4e32-9766-54cbff8053b0-kube-api-access-mppnk\") pod \"55e299a6-a633-4e32-9766-54cbff8053b0\" (UID: \"55e299a6-a633-4e32-9766-54cbff8053b0\") " Dec 09 10:11:03 crc kubenswrapper[4824]: I1209 10:11:03.092486 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55e299a6-a633-4e32-9766-54cbff8053b0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "55e299a6-a633-4e32-9766-54cbff8053b0" (UID: "55e299a6-a633-4e32-9766-54cbff8053b0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:11:03 crc kubenswrapper[4824]: I1209 10:11:03.093026 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55e299a6-a633-4e32-9766-54cbff8053b0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:03 crc kubenswrapper[4824]: I1209 10:11:03.097555 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55e299a6-a633-4e32-9766-54cbff8053b0-kube-api-access-mppnk" (OuterVolumeSpecName: "kube-api-access-mppnk") pod "55e299a6-a633-4e32-9766-54cbff8053b0" (UID: "55e299a6-a633-4e32-9766-54cbff8053b0"). InnerVolumeSpecName "kube-api-access-mppnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:11:03 crc kubenswrapper[4824]: I1209 10:11:03.108270 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nxcqz" Dec 09 10:11:03 crc kubenswrapper[4824]: I1209 10:11:03.108605 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nxcqz" event={"ID":"7ee07f30-f716-4d7a-b5dd-71de0e128d3f","Type":"ContainerDied","Data":"b50356ce8581ade212cd45137fd2c8c55a17a5b1c9a842507228a672e16eff9a"} Dec 09 10:11:03 crc kubenswrapper[4824]: I1209 10:11:03.108875 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b50356ce8581ade212cd45137fd2c8c55a17a5b1c9a842507228a672e16eff9a" Dec 09 10:11:03 crc kubenswrapper[4824]: I1209 10:11:03.114085 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-zh7x5" event={"ID":"55e299a6-a633-4e32-9766-54cbff8053b0","Type":"ContainerDied","Data":"ca4f8974e87ebd586b31f40786a15ddbd93cf215e4425617c4a5342c681451ef"} Dec 09 10:11:03 crc kubenswrapper[4824]: I1209 10:11:03.114178 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca4f8974e87ebd586b31f40786a15ddbd93cf215e4425617c4a5342c681451ef" Dec 09 10:11:03 crc kubenswrapper[4824]: I1209 10:11:03.114275 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-zh7x5" Dec 09 10:11:03 crc kubenswrapper[4824]: I1209 10:11:03.310967 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mppnk\" (UniqueName: \"kubernetes.io/projected/55e299a6-a633-4e32-9766-54cbff8053b0-kube-api-access-mppnk\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.055766 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-9mmtj" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.139835 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0196-account-create-update-llvmc" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.156296 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9px5r" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.169527 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae040f9f-f187-4477-9455-af94eda40a1a-operator-scripts\") pod \"ae040f9f-f187-4477-9455-af94eda40a1a\" (UID: \"ae040f9f-f187-4477-9455-af94eda40a1a\") " Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.169691 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79dm8\" (UniqueName: \"kubernetes.io/projected/ae040f9f-f187-4477-9455-af94eda40a1a-kube-api-access-79dm8\") pod \"ae040f9f-f187-4477-9455-af94eda40a1a\" (UID: \"ae040f9f-f187-4477-9455-af94eda40a1a\") " Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.172075 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae040f9f-f187-4477-9455-af94eda40a1a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ae040f9f-f187-4477-9455-af94eda40a1a" (UID: "ae040f9f-f187-4477-9455-af94eda40a1a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.178333 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae040f9f-f187-4477-9455-af94eda40a1a-kube-api-access-79dm8" (OuterVolumeSpecName: "kube-api-access-79dm8") pod "ae040f9f-f187-4477-9455-af94eda40a1a" (UID: "ae040f9f-f187-4477-9455-af94eda40a1a"). InnerVolumeSpecName "kube-api-access-79dm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.233342 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b08a-account-create-update-jx676" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.240883 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3340-account-create-update-fskfw" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.257646 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-719f-account-create-update-mlf5c" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.274867 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2vp2\" (UniqueName: \"kubernetes.io/projected/826695ab-42b3-496a-a67b-add8ab95d4c1-kube-api-access-p2vp2\") pod \"826695ab-42b3-496a-a67b-add8ab95d4c1\" (UID: \"826695ab-42b3-496a-a67b-add8ab95d4c1\") " Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.274951 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bcqg\" (UniqueName: \"kubernetes.io/projected/f6a3773f-0af2-4536-a361-dd51cedd700f-kube-api-access-9bcqg\") pod \"f6a3773f-0af2-4536-a361-dd51cedd700f\" (UID: \"f6a3773f-0af2-4536-a361-dd51cedd700f\") " Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.275106 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6a3773f-0af2-4536-a361-dd51cedd700f-operator-scripts\") pod \"f6a3773f-0af2-4536-a361-dd51cedd700f\" (UID: \"f6a3773f-0af2-4536-a361-dd51cedd700f\") " Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.275183 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/826695ab-42b3-496a-a67b-add8ab95d4c1-operator-scripts\") pod \"826695ab-42b3-496a-a67b-add8ab95d4c1\" (UID: \"826695ab-42b3-496a-a67b-add8ab95d4c1\") " Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.275879 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae040f9f-f187-4477-9455-af94eda40a1a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.275898 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79dm8\" (UniqueName: \"kubernetes.io/projected/ae040f9f-f187-4477-9455-af94eda40a1a-kube-api-access-79dm8\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.279693 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/826695ab-42b3-496a-a67b-add8ab95d4c1-kube-api-access-p2vp2" (OuterVolumeSpecName: "kube-api-access-p2vp2") pod "826695ab-42b3-496a-a67b-add8ab95d4c1" (UID: "826695ab-42b3-496a-a67b-add8ab95d4c1"). InnerVolumeSpecName "kube-api-access-p2vp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.280200 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6a3773f-0af2-4536-a361-dd51cedd700f-kube-api-access-9bcqg" (OuterVolumeSpecName: "kube-api-access-9bcqg") pod "f6a3773f-0af2-4536-a361-dd51cedd700f" (UID: "f6a3773f-0af2-4536-a361-dd51cedd700f"). InnerVolumeSpecName "kube-api-access-9bcqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.280577 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6a3773f-0af2-4536-a361-dd51cedd700f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f6a3773f-0af2-4536-a361-dd51cedd700f" (UID: "f6a3773f-0af2-4536-a361-dd51cedd700f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.280898 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/826695ab-42b3-496a-a67b-add8ab95d4c1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "826695ab-42b3-496a-a67b-add8ab95d4c1" (UID: "826695ab-42b3-496a-a67b-add8ab95d4c1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.377316 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55a79bdb-6f53-4f93-85cd-e674210cc7e7-operator-scripts\") pod \"55a79bdb-6f53-4f93-85cd-e674210cc7e7\" (UID: \"55a79bdb-6f53-4f93-85cd-e674210cc7e7\") " Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.377395 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2npvb\" (UniqueName: \"kubernetes.io/projected/48b5dcdc-81a1-49f9-b5b6-5c9445be07dc-kube-api-access-2npvb\") pod \"48b5dcdc-81a1-49f9-b5b6-5c9445be07dc\" (UID: \"48b5dcdc-81a1-49f9-b5b6-5c9445be07dc\") " Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.377502 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6tlp\" (UniqueName: \"kubernetes.io/projected/8a2cda53-48d6-480c-9da8-1caec867aee5-kube-api-access-d6tlp\") pod \"8a2cda53-48d6-480c-9da8-1caec867aee5\" (UID: \"8a2cda53-48d6-480c-9da8-1caec867aee5\") " Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.377581 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48b5dcdc-81a1-49f9-b5b6-5c9445be07dc-operator-scripts\") pod \"48b5dcdc-81a1-49f9-b5b6-5c9445be07dc\" (UID: \"48b5dcdc-81a1-49f9-b5b6-5c9445be07dc\") " Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.377666 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5fmv\" (UniqueName: \"kubernetes.io/projected/55a79bdb-6f53-4f93-85cd-e674210cc7e7-kube-api-access-b5fmv\") pod \"55a79bdb-6f53-4f93-85cd-e674210cc7e7\" (UID: \"55a79bdb-6f53-4f93-85cd-e674210cc7e7\") " Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.377772 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a2cda53-48d6-480c-9da8-1caec867aee5-operator-scripts\") pod \"8a2cda53-48d6-480c-9da8-1caec867aee5\" (UID: \"8a2cda53-48d6-480c-9da8-1caec867aee5\") " Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.377982 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55a79bdb-6f53-4f93-85cd-e674210cc7e7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "55a79bdb-6f53-4f93-85cd-e674210cc7e7" (UID: "55a79bdb-6f53-4f93-85cd-e674210cc7e7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.378445 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48b5dcdc-81a1-49f9-b5b6-5c9445be07dc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "48b5dcdc-81a1-49f9-b5b6-5c9445be07dc" (UID: "48b5dcdc-81a1-49f9-b5b6-5c9445be07dc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.378466 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a2cda53-48d6-480c-9da8-1caec867aee5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8a2cda53-48d6-480c-9da8-1caec867aee5" (UID: "8a2cda53-48d6-480c-9da8-1caec867aee5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.379164 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2vp2\" (UniqueName: \"kubernetes.io/projected/826695ab-42b3-496a-a67b-add8ab95d4c1-kube-api-access-p2vp2\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.379189 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bcqg\" (UniqueName: \"kubernetes.io/projected/f6a3773f-0af2-4536-a361-dd51cedd700f-kube-api-access-9bcqg\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.379199 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55a79bdb-6f53-4f93-85cd-e674210cc7e7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.379208 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6a3773f-0af2-4536-a361-dd51cedd700f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.379218 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48b5dcdc-81a1-49f9-b5b6-5c9445be07dc-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.379227 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/826695ab-42b3-496a-a67b-add8ab95d4c1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.379235 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a2cda53-48d6-480c-9da8-1caec867aee5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.381732 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48b5dcdc-81a1-49f9-b5b6-5c9445be07dc-kube-api-access-2npvb" (OuterVolumeSpecName: "kube-api-access-2npvb") pod "48b5dcdc-81a1-49f9-b5b6-5c9445be07dc" (UID: "48b5dcdc-81a1-49f9-b5b6-5c9445be07dc"). InnerVolumeSpecName "kube-api-access-2npvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.382600 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a2cda53-48d6-480c-9da8-1caec867aee5-kube-api-access-d6tlp" (OuterVolumeSpecName: "kube-api-access-d6tlp") pod "8a2cda53-48d6-480c-9da8-1caec867aee5" (UID: "8a2cda53-48d6-480c-9da8-1caec867aee5"). InnerVolumeSpecName "kube-api-access-d6tlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.382975 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55a79bdb-6f53-4f93-85cd-e674210cc7e7-kube-api-access-b5fmv" (OuterVolumeSpecName: "kube-api-access-b5fmv") pod "55a79bdb-6f53-4f93-85cd-e674210cc7e7" (UID: "55a79bdb-6f53-4f93-85cd-e674210cc7e7"). InnerVolumeSpecName "kube-api-access-b5fmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.481172 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6tlp\" (UniqueName: \"kubernetes.io/projected/8a2cda53-48d6-480c-9da8-1caec867aee5-kube-api-access-d6tlp\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.481210 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5fmv\" (UniqueName: \"kubernetes.io/projected/55a79bdb-6f53-4f93-85cd-e674210cc7e7-kube-api-access-b5fmv\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.481223 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2npvb\" (UniqueName: \"kubernetes.io/projected/48b5dcdc-81a1-49f9-b5b6-5c9445be07dc-kube-api-access-2npvb\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.482144 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9px5r" event={"ID":"f6a3773f-0af2-4536-a361-dd51cedd700f","Type":"ContainerDied","Data":"aff45440dcdeb5478ac01bebab61b3cd4b87b1ce6ff500a4014306a6508f4c8c"} Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.482268 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aff45440dcdeb5478ac01bebab61b3cd4b87b1ce6ff500a4014306a6508f4c8c" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.482182 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9px5r" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.487080 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0196-account-create-update-llvmc" event={"ID":"826695ab-42b3-496a-a67b-add8ab95d4c1","Type":"ContainerDied","Data":"d52c8e2c2279b8e82884a2f9c7911d242003cfb8fd1fd30a45f0ff04593bff5a"} Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.487115 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d52c8e2c2279b8e82884a2f9c7911d242003cfb8fd1fd30a45f0ff04593bff5a" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.487172 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0196-account-create-update-llvmc" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.490891 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b08a-account-create-update-jx676" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.491057 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b08a-account-create-update-jx676" event={"ID":"55a79bdb-6f53-4f93-85cd-e674210cc7e7","Type":"ContainerDied","Data":"7b26c83199062e59c7c027ef919d3888780d2e59210d952ea41bfe241f5717ff"} Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.491145 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b26c83199062e59c7c027ef919d3888780d2e59210d952ea41bfe241f5717ff" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.493129 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-6swf9" event={"ID":"da292326-c56e-42d0-bfe2-2b75b53c210d","Type":"ContainerStarted","Data":"a6aae0cc2f2d73cc679c1e9ba8e70322c362a1997846b0fadb6d5a23a5230277"} Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.498824 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-9mmtj" event={"ID":"ae040f9f-f187-4477-9455-af94eda40a1a","Type":"ContainerDied","Data":"6f1be4155832bc397848c30ec15e85fddfb2236173335068afc14bf85853c0f5"} Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.498874 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f1be4155832bc397848c30ec15e85fddfb2236173335068afc14bf85853c0f5" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.498944 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-9mmtj" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.501718 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-719f-account-create-update-mlf5c" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.501748 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-719f-account-create-update-mlf5c" event={"ID":"48b5dcdc-81a1-49f9-b5b6-5c9445be07dc","Type":"ContainerDied","Data":"d42df4baf36fb684468cd910ef6fabd3db82bc040a61d91edb8840eafc9f1ce1"} Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.501798 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d42df4baf36fb684468cd910ef6fabd3db82bc040a61d91edb8840eafc9f1ce1" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.503907 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3340-account-create-update-fskfw" event={"ID":"8a2cda53-48d6-480c-9da8-1caec867aee5","Type":"ContainerDied","Data":"dafa8bf8cf3a340f7ac03ec0b5a7990b5fa9c89716729d5c7b9820fab179799e"} Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.503958 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dafa8bf8cf3a340f7ac03ec0b5a7990b5fa9c89716729d5c7b9820fab179799e" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.503938 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3340-account-create-update-fskfw" Dec 09 10:11:07 crc kubenswrapper[4824]: I1209 10:11:07.521321 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-6swf9" podStartSLOduration=3.809824981 podStartE2EDuration="10.521295603s" podCreationTimestamp="2025-12-09 10:10:57 +0000 UTC" firstStartedPulling="2025-12-09 10:11:00.246326915 +0000 UTC m=+1416.580831582" lastFinishedPulling="2025-12-09 10:11:06.957797537 +0000 UTC m=+1423.292302204" observedRunningTime="2025-12-09 10:11:07.50801948 +0000 UTC m=+1423.842524167" watchObservedRunningTime="2025-12-09 10:11:07.521295603 +0000 UTC m=+1423.855800270" Dec 09 10:11:11 crc kubenswrapper[4824]: I1209 10:11:11.550075 4824 generic.go:334] "Generic (PLEG): container finished" podID="da292326-c56e-42d0-bfe2-2b75b53c210d" containerID="a6aae0cc2f2d73cc679c1e9ba8e70322c362a1997846b0fadb6d5a23a5230277" exitCode=0 Dec 09 10:11:11 crc kubenswrapper[4824]: I1209 10:11:11.550183 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-6swf9" event={"ID":"da292326-c56e-42d0-bfe2-2b75b53c210d","Type":"ContainerDied","Data":"a6aae0cc2f2d73cc679c1e9ba8e70322c362a1997846b0fadb6d5a23a5230277"} Dec 09 10:11:12 crc kubenswrapper[4824]: I1209 10:11:12.891927 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-6swf9" Dec 09 10:11:12 crc kubenswrapper[4824]: I1209 10:11:12.919706 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da292326-c56e-42d0-bfe2-2b75b53c210d-combined-ca-bundle\") pod \"da292326-c56e-42d0-bfe2-2b75b53c210d\" (UID: \"da292326-c56e-42d0-bfe2-2b75b53c210d\") " Dec 09 10:11:12 crc kubenswrapper[4824]: I1209 10:11:12.919774 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7w82g\" (UniqueName: \"kubernetes.io/projected/da292326-c56e-42d0-bfe2-2b75b53c210d-kube-api-access-7w82g\") pod \"da292326-c56e-42d0-bfe2-2b75b53c210d\" (UID: \"da292326-c56e-42d0-bfe2-2b75b53c210d\") " Dec 09 10:11:12 crc kubenswrapper[4824]: I1209 10:11:12.919811 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da292326-c56e-42d0-bfe2-2b75b53c210d-config-data\") pod \"da292326-c56e-42d0-bfe2-2b75b53c210d\" (UID: \"da292326-c56e-42d0-bfe2-2b75b53c210d\") " Dec 09 10:11:12 crc kubenswrapper[4824]: I1209 10:11:12.926079 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da292326-c56e-42d0-bfe2-2b75b53c210d-kube-api-access-7w82g" (OuterVolumeSpecName: "kube-api-access-7w82g") pod "da292326-c56e-42d0-bfe2-2b75b53c210d" (UID: "da292326-c56e-42d0-bfe2-2b75b53c210d"). InnerVolumeSpecName "kube-api-access-7w82g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:11:12 crc kubenswrapper[4824]: I1209 10:11:12.990017 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da292326-c56e-42d0-bfe2-2b75b53c210d-config-data" (OuterVolumeSpecName: "config-data") pod "da292326-c56e-42d0-bfe2-2b75b53c210d" (UID: "da292326-c56e-42d0-bfe2-2b75b53c210d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:11:12 crc kubenswrapper[4824]: I1209 10:11:12.990501 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da292326-c56e-42d0-bfe2-2b75b53c210d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "da292326-c56e-42d0-bfe2-2b75b53c210d" (UID: "da292326-c56e-42d0-bfe2-2b75b53c210d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.022455 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da292326-c56e-42d0-bfe2-2b75b53c210d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.022493 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7w82g\" (UniqueName: \"kubernetes.io/projected/da292326-c56e-42d0-bfe2-2b75b53c210d-kube-api-access-7w82g\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.022508 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da292326-c56e-42d0-bfe2-2b75b53c210d-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.572912 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-6swf9" event={"ID":"da292326-c56e-42d0-bfe2-2b75b53c210d","Type":"ContainerDied","Data":"70ac9f5b948a7c677d0c3ad55d67442affab350383f6b2aa57006ae9932329d5"} Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.572981 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70ac9f5b948a7c677d0c3ad55d67442affab350383f6b2aa57006ae9932329d5" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.572941 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-6swf9" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.899444 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-st484"] Dec 09 10:11:13 crc kubenswrapper[4824]: E1209 10:11:13.900403 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55a79bdb-6f53-4f93-85cd-e674210cc7e7" containerName="mariadb-account-create-update" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.900421 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="55a79bdb-6f53-4f93-85cd-e674210cc7e7" containerName="mariadb-account-create-update" Dec 09 10:11:13 crc kubenswrapper[4824]: E1209 10:11:13.900431 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6a3773f-0af2-4536-a361-dd51cedd700f" containerName="mariadb-database-create" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.900437 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6a3773f-0af2-4536-a361-dd51cedd700f" containerName="mariadb-database-create" Dec 09 10:11:13 crc kubenswrapper[4824]: E1209 10:11:13.900448 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="826695ab-42b3-496a-a67b-add8ab95d4c1" containerName="mariadb-account-create-update" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.900455 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="826695ab-42b3-496a-a67b-add8ab95d4c1" containerName="mariadb-account-create-update" Dec 09 10:11:13 crc kubenswrapper[4824]: E1209 10:11:13.900467 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ee07f30-f716-4d7a-b5dd-71de0e128d3f" containerName="mariadb-database-create" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.900475 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ee07f30-f716-4d7a-b5dd-71de0e128d3f" containerName="mariadb-database-create" Dec 09 10:11:13 crc kubenswrapper[4824]: E1209 10:11:13.900486 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55e299a6-a633-4e32-9766-54cbff8053b0" containerName="mariadb-database-create" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.900492 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="55e299a6-a633-4e32-9766-54cbff8053b0" containerName="mariadb-database-create" Dec 09 10:11:13 crc kubenswrapper[4824]: E1209 10:11:13.900500 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a2cda53-48d6-480c-9da8-1caec867aee5" containerName="mariadb-account-create-update" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.900506 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a2cda53-48d6-480c-9da8-1caec867aee5" containerName="mariadb-account-create-update" Dec 09 10:11:13 crc kubenswrapper[4824]: E1209 10:11:13.900543 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da292326-c56e-42d0-bfe2-2b75b53c210d" containerName="keystone-db-sync" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.900549 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="da292326-c56e-42d0-bfe2-2b75b53c210d" containerName="keystone-db-sync" Dec 09 10:11:13 crc kubenswrapper[4824]: E1209 10:11:13.900559 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48b5dcdc-81a1-49f9-b5b6-5c9445be07dc" containerName="mariadb-account-create-update" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.900565 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="48b5dcdc-81a1-49f9-b5b6-5c9445be07dc" containerName="mariadb-account-create-update" Dec 09 10:11:13 crc kubenswrapper[4824]: E1209 10:11:13.900577 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae040f9f-f187-4477-9455-af94eda40a1a" containerName="mariadb-database-create" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.900583 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae040f9f-f187-4477-9455-af94eda40a1a" containerName="mariadb-database-create" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.900825 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae040f9f-f187-4477-9455-af94eda40a1a" containerName="mariadb-database-create" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.900840 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="55a79bdb-6f53-4f93-85cd-e674210cc7e7" containerName="mariadb-account-create-update" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.900856 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="55e299a6-a633-4e32-9766-54cbff8053b0" containerName="mariadb-database-create" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.900864 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a2cda53-48d6-480c-9da8-1caec867aee5" containerName="mariadb-account-create-update" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.900881 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="da292326-c56e-42d0-bfe2-2b75b53c210d" containerName="keystone-db-sync" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.900888 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="826695ab-42b3-496a-a67b-add8ab95d4c1" containerName="mariadb-account-create-update" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.900896 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="48b5dcdc-81a1-49f9-b5b6-5c9445be07dc" containerName="mariadb-account-create-update" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.900908 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ee07f30-f716-4d7a-b5dd-71de0e128d3f" containerName="mariadb-database-create" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.900929 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6a3773f-0af2-4536-a361-dd51cedd700f" containerName="mariadb-database-create" Dec 09 10:11:13 crc kubenswrapper[4824]: I1209 10:11:13.904634 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-st484" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.012457 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-st484"] Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.065575 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-st484\" (UID: \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\") " pod="openstack/dnsmasq-dns-847c4cc679-st484" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.065860 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-dns-svc\") pod \"dnsmasq-dns-847c4cc679-st484\" (UID: \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\") " pod="openstack/dnsmasq-dns-847c4cc679-st484" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.065888 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tfhn\" (UniqueName: \"kubernetes.io/projected/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-kube-api-access-9tfhn\") pod \"dnsmasq-dns-847c4cc679-st484\" (UID: \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\") " pod="openstack/dnsmasq-dns-847c4cc679-st484" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.066015 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-st484\" (UID: \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\") " pod="openstack/dnsmasq-dns-847c4cc679-st484" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.066076 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-config\") pod \"dnsmasq-dns-847c4cc679-st484\" (UID: \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\") " pod="openstack/dnsmasq-dns-847c4cc679-st484" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.066140 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-st484\" (UID: \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\") " pod="openstack/dnsmasq-dns-847c4cc679-st484" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.075889 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-6lhrx"] Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.089083 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6lhrx" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.101985 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ggbf4" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.102315 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.102317 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.108161 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.316501 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.337136 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-st484\" (UID: \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\") " pod="openstack/dnsmasq-dns-847c4cc679-st484" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.340876 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-config\") pod \"dnsmasq-dns-847c4cc679-st484\" (UID: \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\") " pod="openstack/dnsmasq-dns-847c4cc679-st484" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.341028 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-fernet-keys\") pod \"keystone-bootstrap-6lhrx\" (UID: \"c7e2a035-7568-4455-a3bb-329dcf3276ca\") " pod="openstack/keystone-bootstrap-6lhrx" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.341152 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-st484\" (UID: \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\") " pod="openstack/dnsmasq-dns-847c4cc679-st484" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.341224 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-scripts\") pod \"keystone-bootstrap-6lhrx\" (UID: \"c7e2a035-7568-4455-a3bb-329dcf3276ca\") " pod="openstack/keystone-bootstrap-6lhrx" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.341253 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwhzk\" (UniqueName: \"kubernetes.io/projected/c7e2a035-7568-4455-a3bb-329dcf3276ca-kube-api-access-kwhzk\") pod \"keystone-bootstrap-6lhrx\" (UID: \"c7e2a035-7568-4455-a3bb-329dcf3276ca\") " pod="openstack/keystone-bootstrap-6lhrx" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.341599 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-config-data\") pod \"keystone-bootstrap-6lhrx\" (UID: \"c7e2a035-7568-4455-a3bb-329dcf3276ca\") " pod="openstack/keystone-bootstrap-6lhrx" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.341703 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-st484\" (UID: \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\") " pod="openstack/dnsmasq-dns-847c4cc679-st484" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.341838 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-combined-ca-bundle\") pod \"keystone-bootstrap-6lhrx\" (UID: \"c7e2a035-7568-4455-a3bb-329dcf3276ca\") " pod="openstack/keystone-bootstrap-6lhrx" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.342011 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-credential-keys\") pod \"keystone-bootstrap-6lhrx\" (UID: \"c7e2a035-7568-4455-a3bb-329dcf3276ca\") " pod="openstack/keystone-bootstrap-6lhrx" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.342078 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-dns-svc\") pod \"dnsmasq-dns-847c4cc679-st484\" (UID: \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\") " pod="openstack/dnsmasq-dns-847c4cc679-st484" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.342101 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tfhn\" (UniqueName: \"kubernetes.io/projected/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-kube-api-access-9tfhn\") pod \"dnsmasq-dns-847c4cc679-st484\" (UID: \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\") " pod="openstack/dnsmasq-dns-847c4cc679-st484" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.343825 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-st484\" (UID: \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\") " pod="openstack/dnsmasq-dns-847c4cc679-st484" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.345571 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-st484\" (UID: \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\") " pod="openstack/dnsmasq-dns-847c4cc679-st484" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.346332 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-st484\" (UID: \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\") " pod="openstack/dnsmasq-dns-847c4cc679-st484" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.347060 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-dns-svc\") pod \"dnsmasq-dns-847c4cc679-st484\" (UID: \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\") " pod="openstack/dnsmasq-dns-847c4cc679-st484" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.347170 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6lhrx"] Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.364870 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-config\") pod \"dnsmasq-dns-847c4cc679-st484\" (UID: \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\") " pod="openstack/dnsmasq-dns-847c4cc679-st484" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.394133 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tfhn\" (UniqueName: \"kubernetes.io/projected/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-kube-api-access-9tfhn\") pod \"dnsmasq-dns-847c4cc679-st484\" (UID: \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\") " pod="openstack/dnsmasq-dns-847c4cc679-st484" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.399516 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-vfnsw"] Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.405942 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-vfnsw" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.438261 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-vfnsw"] Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.441870 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-k8rrz" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.442208 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.444971 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-credential-keys\") pod \"keystone-bootstrap-6lhrx\" (UID: \"c7e2a035-7568-4455-a3bb-329dcf3276ca\") " pod="openstack/keystone-bootstrap-6lhrx" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.445097 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-fernet-keys\") pod \"keystone-bootstrap-6lhrx\" (UID: \"c7e2a035-7568-4455-a3bb-329dcf3276ca\") " pod="openstack/keystone-bootstrap-6lhrx" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.445143 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-scripts\") pod \"keystone-bootstrap-6lhrx\" (UID: \"c7e2a035-7568-4455-a3bb-329dcf3276ca\") " pod="openstack/keystone-bootstrap-6lhrx" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.445166 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwhzk\" (UniqueName: \"kubernetes.io/projected/c7e2a035-7568-4455-a3bb-329dcf3276ca-kube-api-access-kwhzk\") pod \"keystone-bootstrap-6lhrx\" (UID: \"c7e2a035-7568-4455-a3bb-329dcf3276ca\") " pod="openstack/keystone-bootstrap-6lhrx" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.445269 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-config-data\") pod \"keystone-bootstrap-6lhrx\" (UID: \"c7e2a035-7568-4455-a3bb-329dcf3276ca\") " pod="openstack/keystone-bootstrap-6lhrx" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.445329 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-combined-ca-bundle\") pod \"keystone-bootstrap-6lhrx\" (UID: \"c7e2a035-7568-4455-a3bb-329dcf3276ca\") " pod="openstack/keystone-bootstrap-6lhrx" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.457825 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-fernet-keys\") pod \"keystone-bootstrap-6lhrx\" (UID: \"c7e2a035-7568-4455-a3bb-329dcf3276ca\") " pod="openstack/keystone-bootstrap-6lhrx" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.459701 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-config-data\") pod \"keystone-bootstrap-6lhrx\" (UID: \"c7e2a035-7568-4455-a3bb-329dcf3276ca\") " pod="openstack/keystone-bootstrap-6lhrx" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.468248 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-scripts\") pod \"keystone-bootstrap-6lhrx\" (UID: \"c7e2a035-7568-4455-a3bb-329dcf3276ca\") " pod="openstack/keystone-bootstrap-6lhrx" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.490505 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-credential-keys\") pod \"keystone-bootstrap-6lhrx\" (UID: \"c7e2a035-7568-4455-a3bb-329dcf3276ca\") " pod="openstack/keystone-bootstrap-6lhrx" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.495452 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-combined-ca-bundle\") pod \"keystone-bootstrap-6lhrx\" (UID: \"c7e2a035-7568-4455-a3bb-329dcf3276ca\") " pod="openstack/keystone-bootstrap-6lhrx" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.505910 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-mqjtv"] Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.508024 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mqjtv" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.510382 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwhzk\" (UniqueName: \"kubernetes.io/projected/c7e2a035-7568-4455-a3bb-329dcf3276ca-kube-api-access-kwhzk\") pod \"keystone-bootstrap-6lhrx\" (UID: \"c7e2a035-7568-4455-a3bb-329dcf3276ca\") " pod="openstack/keystone-bootstrap-6lhrx" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.533189 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.533580 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.533743 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-r4m6h" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.546245 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-d4klr"] Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.547565 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2csn\" (UniqueName: \"kubernetes.io/projected/405d0a5e-fd62-4146-bfb9-96ff6c077836-kube-api-access-p2csn\") pod \"heat-db-sync-vfnsw\" (UID: \"405d0a5e-fd62-4146-bfb9-96ff6c077836\") " pod="openstack/heat-db-sync-vfnsw" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.547750 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/405d0a5e-fd62-4146-bfb9-96ff6c077836-combined-ca-bundle\") pod \"heat-db-sync-vfnsw\" (UID: \"405d0a5e-fd62-4146-bfb9-96ff6c077836\") " pod="openstack/heat-db-sync-vfnsw" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.547873 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/405d0a5e-fd62-4146-bfb9-96ff6c077836-config-data\") pod \"heat-db-sync-vfnsw\" (UID: \"405d0a5e-fd62-4146-bfb9-96ff6c077836\") " pod="openstack/heat-db-sync-vfnsw" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.548634 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-d4klr" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.561912 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-st484" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.607993 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-mqjtv"] Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.723574 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6lhrx" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.724869 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3d076df-d1e1-473e-9584-79b30cb6d023-combined-ca-bundle\") pod \"barbican-db-sync-d4klr\" (UID: \"d3d076df-d1e1-473e-9584-79b30cb6d023\") " pod="openstack/barbican-db-sync-d4klr" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.724920 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2csn\" (UniqueName: \"kubernetes.io/projected/405d0a5e-fd62-4146-bfb9-96ff6c077836-kube-api-access-p2csn\") pod \"heat-db-sync-vfnsw\" (UID: \"405d0a5e-fd62-4146-bfb9-96ff6c077836\") " pod="openstack/heat-db-sync-vfnsw" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.724967 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh2p2\" (UniqueName: \"kubernetes.io/projected/d3d076df-d1e1-473e-9584-79b30cb6d023-kube-api-access-mh2p2\") pod \"barbican-db-sync-d4klr\" (UID: \"d3d076df-d1e1-473e-9584-79b30cb6d023\") " pod="openstack/barbican-db-sync-d4klr" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.724998 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr4gr\" (UniqueName: \"kubernetes.io/projected/8eb90797-2763-4ffb-8fa5-32a1c8db5c84-kube-api-access-rr4gr\") pod \"neutron-db-sync-mqjtv\" (UID: \"8eb90797-2763-4ffb-8fa5-32a1c8db5c84\") " pod="openstack/neutron-db-sync-mqjtv" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.725035 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3d076df-d1e1-473e-9584-79b30cb6d023-db-sync-config-data\") pod \"barbican-db-sync-d4klr\" (UID: \"d3d076df-d1e1-473e-9584-79b30cb6d023\") " pod="openstack/barbican-db-sync-d4klr" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.725136 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/405d0a5e-fd62-4146-bfb9-96ff6c077836-combined-ca-bundle\") pod \"heat-db-sync-vfnsw\" (UID: \"405d0a5e-fd62-4146-bfb9-96ff6c077836\") " pod="openstack/heat-db-sync-vfnsw" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.725247 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/405d0a5e-fd62-4146-bfb9-96ff6c077836-config-data\") pod \"heat-db-sync-vfnsw\" (UID: \"405d0a5e-fd62-4146-bfb9-96ff6c077836\") " pod="openstack/heat-db-sync-vfnsw" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.725285 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8eb90797-2763-4ffb-8fa5-32a1c8db5c84-combined-ca-bundle\") pod \"neutron-db-sync-mqjtv\" (UID: \"8eb90797-2763-4ffb-8fa5-32a1c8db5c84\") " pod="openstack/neutron-db-sync-mqjtv" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.725337 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8eb90797-2763-4ffb-8fa5-32a1c8db5c84-config\") pod \"neutron-db-sync-mqjtv\" (UID: \"8eb90797-2763-4ffb-8fa5-32a1c8db5c84\") " pod="openstack/neutron-db-sync-mqjtv" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.729466 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/405d0a5e-fd62-4146-bfb9-96ff6c077836-config-data\") pod \"heat-db-sync-vfnsw\" (UID: \"405d0a5e-fd62-4146-bfb9-96ff6c077836\") " pod="openstack/heat-db-sync-vfnsw" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.745467 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/405d0a5e-fd62-4146-bfb9-96ff6c077836-combined-ca-bundle\") pod \"heat-db-sync-vfnsw\" (UID: \"405d0a5e-fd62-4146-bfb9-96ff6c077836\") " pod="openstack/heat-db-sync-vfnsw" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.766152 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.766432 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-gdb54" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.783501 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-d4klr"] Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.804960 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2csn\" (UniqueName: \"kubernetes.io/projected/405d0a5e-fd62-4146-bfb9-96ff6c077836-kube-api-access-p2csn\") pod \"heat-db-sync-vfnsw\" (UID: \"405d0a5e-fd62-4146-bfb9-96ff6c077836\") " pod="openstack/heat-db-sync-vfnsw" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.826638 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3d076df-d1e1-473e-9584-79b30cb6d023-combined-ca-bundle\") pod \"barbican-db-sync-d4klr\" (UID: \"d3d076df-d1e1-473e-9584-79b30cb6d023\") " pod="openstack/barbican-db-sync-d4klr" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.826756 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh2p2\" (UniqueName: \"kubernetes.io/projected/d3d076df-d1e1-473e-9584-79b30cb6d023-kube-api-access-mh2p2\") pod \"barbican-db-sync-d4klr\" (UID: \"d3d076df-d1e1-473e-9584-79b30cb6d023\") " pod="openstack/barbican-db-sync-d4klr" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.826824 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr4gr\" (UniqueName: \"kubernetes.io/projected/8eb90797-2763-4ffb-8fa5-32a1c8db5c84-kube-api-access-rr4gr\") pod \"neutron-db-sync-mqjtv\" (UID: \"8eb90797-2763-4ffb-8fa5-32a1c8db5c84\") " pod="openstack/neutron-db-sync-mqjtv" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.826872 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3d076df-d1e1-473e-9584-79b30cb6d023-db-sync-config-data\") pod \"barbican-db-sync-d4klr\" (UID: \"d3d076df-d1e1-473e-9584-79b30cb6d023\") " pod="openstack/barbican-db-sync-d4klr" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.826987 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8eb90797-2763-4ffb-8fa5-32a1c8db5c84-combined-ca-bundle\") pod \"neutron-db-sync-mqjtv\" (UID: \"8eb90797-2763-4ffb-8fa5-32a1c8db5c84\") " pod="openstack/neutron-db-sync-mqjtv" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.827026 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8eb90797-2763-4ffb-8fa5-32a1c8db5c84-config\") pod \"neutron-db-sync-mqjtv\" (UID: \"8eb90797-2763-4ffb-8fa5-32a1c8db5c84\") " pod="openstack/neutron-db-sync-mqjtv" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.846982 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8eb90797-2763-4ffb-8fa5-32a1c8db5c84-combined-ca-bundle\") pod \"neutron-db-sync-mqjtv\" (UID: \"8eb90797-2763-4ffb-8fa5-32a1c8db5c84\") " pod="openstack/neutron-db-sync-mqjtv" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.848084 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3d076df-d1e1-473e-9584-79b30cb6d023-db-sync-config-data\") pod \"barbican-db-sync-d4klr\" (UID: \"d3d076df-d1e1-473e-9584-79b30cb6d023\") " pod="openstack/barbican-db-sync-d4klr" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.849157 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8eb90797-2763-4ffb-8fa5-32a1c8db5c84-config\") pod \"neutron-db-sync-mqjtv\" (UID: \"8eb90797-2763-4ffb-8fa5-32a1c8db5c84\") " pod="openstack/neutron-db-sync-mqjtv" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.851186 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3d076df-d1e1-473e-9584-79b30cb6d023-combined-ca-bundle\") pod \"barbican-db-sync-d4klr\" (UID: \"d3d076df-d1e1-473e-9584-79b30cb6d023\") " pod="openstack/barbican-db-sync-d4klr" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.863362 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-jghjq"] Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.865407 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-jghjq" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.901514 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.901750 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.904305 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh2p2\" (UniqueName: \"kubernetes.io/projected/d3d076df-d1e1-473e-9584-79b30cb6d023-kube-api-access-mh2p2\") pod \"barbican-db-sync-d4klr\" (UID: \"d3d076df-d1e1-473e-9584-79b30cb6d023\") " pod="openstack/barbican-db-sync-d4klr" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.904807 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-nzl6v" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.904951 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-vfnsw" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.906517 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr4gr\" (UniqueName: \"kubernetes.io/projected/8eb90797-2763-4ffb-8fa5-32a1c8db5c84-kube-api-access-rr4gr\") pod \"neutron-db-sync-mqjtv\" (UID: \"8eb90797-2763-4ffb-8fa5-32a1c8db5c84\") " pod="openstack/neutron-db-sync-mqjtv" Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.958005 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-jghjq"] Dec 09 10:11:14 crc kubenswrapper[4824]: I1209 10:11:14.968045 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mqjtv" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.182287 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwnjz\" (UniqueName: \"kubernetes.io/projected/dbceafb7-272a-45fc-9b56-df631986f90a-kube-api-access-lwnjz\") pod \"cinder-db-sync-jghjq\" (UID: \"dbceafb7-272a-45fc-9b56-df631986f90a\") " pod="openstack/cinder-db-sync-jghjq" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.182668 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbceafb7-272a-45fc-9b56-df631986f90a-scripts\") pod \"cinder-db-sync-jghjq\" (UID: \"dbceafb7-272a-45fc-9b56-df631986f90a\") " pod="openstack/cinder-db-sync-jghjq" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.182714 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dbceafb7-272a-45fc-9b56-df631986f90a-db-sync-config-data\") pod \"cinder-db-sync-jghjq\" (UID: \"dbceafb7-272a-45fc-9b56-df631986f90a\") " pod="openstack/cinder-db-sync-jghjq" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.182901 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbceafb7-272a-45fc-9b56-df631986f90a-config-data\") pod \"cinder-db-sync-jghjq\" (UID: \"dbceafb7-272a-45fc-9b56-df631986f90a\") " pod="openstack/cinder-db-sync-jghjq" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.182925 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dbceafb7-272a-45fc-9b56-df631986f90a-etc-machine-id\") pod \"cinder-db-sync-jghjq\" (UID: \"dbceafb7-272a-45fc-9b56-df631986f90a\") " pod="openstack/cinder-db-sync-jghjq" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.182940 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbceafb7-272a-45fc-9b56-df631986f90a-combined-ca-bundle\") pod \"cinder-db-sync-jghjq\" (UID: \"dbceafb7-272a-45fc-9b56-df631986f90a\") " pod="openstack/cinder-db-sync-jghjq" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.233906 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-d4klr" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.291223 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwnjz\" (UniqueName: \"kubernetes.io/projected/dbceafb7-272a-45fc-9b56-df631986f90a-kube-api-access-lwnjz\") pod \"cinder-db-sync-jghjq\" (UID: \"dbceafb7-272a-45fc-9b56-df631986f90a\") " pod="openstack/cinder-db-sync-jghjq" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.291289 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbceafb7-272a-45fc-9b56-df631986f90a-scripts\") pod \"cinder-db-sync-jghjq\" (UID: \"dbceafb7-272a-45fc-9b56-df631986f90a\") " pod="openstack/cinder-db-sync-jghjq" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.291357 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dbceafb7-272a-45fc-9b56-df631986f90a-db-sync-config-data\") pod \"cinder-db-sync-jghjq\" (UID: \"dbceafb7-272a-45fc-9b56-df631986f90a\") " pod="openstack/cinder-db-sync-jghjq" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.291586 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbceafb7-272a-45fc-9b56-df631986f90a-config-data\") pod \"cinder-db-sync-jghjq\" (UID: \"dbceafb7-272a-45fc-9b56-df631986f90a\") " pod="openstack/cinder-db-sync-jghjq" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.291622 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dbceafb7-272a-45fc-9b56-df631986f90a-etc-machine-id\") pod \"cinder-db-sync-jghjq\" (UID: \"dbceafb7-272a-45fc-9b56-df631986f90a\") " pod="openstack/cinder-db-sync-jghjq" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.291643 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbceafb7-272a-45fc-9b56-df631986f90a-combined-ca-bundle\") pod \"cinder-db-sync-jghjq\" (UID: \"dbceafb7-272a-45fc-9b56-df631986f90a\") " pod="openstack/cinder-db-sync-jghjq" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.293963 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dbceafb7-272a-45fc-9b56-df631986f90a-etc-machine-id\") pod \"cinder-db-sync-jghjq\" (UID: \"dbceafb7-272a-45fc-9b56-df631986f90a\") " pod="openstack/cinder-db-sync-jghjq" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.302144 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dbceafb7-272a-45fc-9b56-df631986f90a-db-sync-config-data\") pod \"cinder-db-sync-jghjq\" (UID: \"dbceafb7-272a-45fc-9b56-df631986f90a\") " pod="openstack/cinder-db-sync-jghjq" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.302954 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbceafb7-272a-45fc-9b56-df631986f90a-scripts\") pod \"cinder-db-sync-jghjq\" (UID: \"dbceafb7-272a-45fc-9b56-df631986f90a\") " pod="openstack/cinder-db-sync-jghjq" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.309332 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbceafb7-272a-45fc-9b56-df631986f90a-config-data\") pod \"cinder-db-sync-jghjq\" (UID: \"dbceafb7-272a-45fc-9b56-df631986f90a\") " pod="openstack/cinder-db-sync-jghjq" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.311050 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbceafb7-272a-45fc-9b56-df631986f90a-combined-ca-bundle\") pod \"cinder-db-sync-jghjq\" (UID: \"dbceafb7-272a-45fc-9b56-df631986f90a\") " pod="openstack/cinder-db-sync-jghjq" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.326676 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-st484"] Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.337682 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwnjz\" (UniqueName: \"kubernetes.io/projected/dbceafb7-272a-45fc-9b56-df631986f90a-kube-api-access-lwnjz\") pod \"cinder-db-sync-jghjq\" (UID: \"dbceafb7-272a-45fc-9b56-df631986f90a\") " pod="openstack/cinder-db-sync-jghjq" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.585002 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.585730 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-jghjq" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.587145 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.608377 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.610489 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-vpbz2" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.638671 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.639019 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.645122 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.659250 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.661607 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.692589 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.693054 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.790657 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-ql4lt"] Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.793278 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ql4lt" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.803445 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxs6n\" (UniqueName: \"kubernetes.io/projected/45021af5-224b-4139-a7b6-06ce097681f4-kube-api-access-lxs6n\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.803526 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45021af5-224b-4139-a7b6-06ce097681f4-logs\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.803583 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/45021af5-224b-4139-a7b6-06ce097681f4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.803634 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45021af5-224b-4139-a7b6-06ce097681f4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.803665 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45021af5-224b-4139-a7b6-06ce097681f4-config-data\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.803730 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45021af5-224b-4139-a7b6-06ce097681f4-scripts\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.803822 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.803894 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45021af5-224b-4139-a7b6-06ce097681f4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.804647 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-9snwp" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.804862 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.809304 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 09 10:11:15 crc kubenswrapper[4824]: I1209 10:11:15.827044 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.081390 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.081453 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aae4e0f1-0295-4964-8b34-56bdab58a152-logs\") pod \"placement-db-sync-ql4lt\" (UID: \"aae4e0f1-0295-4964-8b34-56bdab58a152\") " pod="openstack/placement-db-sync-ql4lt" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.081491 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxs6n\" (UniqueName: \"kubernetes.io/projected/45021af5-224b-4139-a7b6-06ce097681f4-kube-api-access-lxs6n\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.081672 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45021af5-224b-4139-a7b6-06ce097681f4-logs\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.081873 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-logs\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.081911 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/45021af5-224b-4139-a7b6-06ce097681f4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.081986 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aae4e0f1-0295-4964-8b34-56bdab58a152-config-data\") pod \"placement-db-sync-ql4lt\" (UID: \"aae4e0f1-0295-4964-8b34-56bdab58a152\") " pod="openstack/placement-db-sync-ql4lt" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.082017 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-57rtp"] Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.082083 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt9zl\" (UniqueName: \"kubernetes.io/projected/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-kube-api-access-mt9zl\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.082128 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45021af5-224b-4139-a7b6-06ce097681f4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.082158 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz2wp\" (UniqueName: \"kubernetes.io/projected/aae4e0f1-0295-4964-8b34-56bdab58a152-kube-api-access-tz2wp\") pod \"placement-db-sync-ql4lt\" (UID: \"aae4e0f1-0295-4964-8b34-56bdab58a152\") " pod="openstack/placement-db-sync-ql4lt" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.082193 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45021af5-224b-4139-a7b6-06ce097681f4-config-data\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.082235 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.082268 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aae4e0f1-0295-4964-8b34-56bdab58a152-combined-ca-bundle\") pod \"placement-db-sync-ql4lt\" (UID: \"aae4e0f1-0295-4964-8b34-56bdab58a152\") " pod="openstack/placement-db-sync-ql4lt" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.082404 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45021af5-224b-4139-a7b6-06ce097681f4-logs\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.082427 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45021af5-224b-4139-a7b6-06ce097681f4-scripts\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.082499 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aae4e0f1-0295-4964-8b34-56bdab58a152-scripts\") pod \"placement-db-sync-ql4lt\" (UID: \"aae4e0f1-0295-4964-8b34-56bdab58a152\") " pod="openstack/placement-db-sync-ql4lt" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.082536 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.082691 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.082721 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.082892 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.082967 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.083031 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45021af5-224b-4139-a7b6-06ce097681f4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.085265 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/45021af5-224b-4139-a7b6-06ce097681f4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.086851 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.091969 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.097998 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45021af5-224b-4139-a7b6-06ce097681f4-scripts\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.098286 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45021af5-224b-4139-a7b6-06ce097681f4-config-data\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.114181 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-ql4lt"] Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.115444 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45021af5-224b-4139-a7b6-06ce097681f4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.118890 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.121416 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.122155 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.123423 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.123461 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/afe41c6e88a0ab9650f976db82a109be2feb044bddc024f73824c502bea72164/globalmount\"" pod="openstack/glance-default-external-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.152289 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.198242 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz2wp\" (UniqueName: \"kubernetes.io/projected/aae4e0f1-0295-4964-8b34-56bdab58a152-kube-api-access-tz2wp\") pod \"placement-db-sync-ql4lt\" (UID: \"aae4e0f1-0295-4964-8b34-56bdab58a152\") " pod="openstack/placement-db-sync-ql4lt" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.198358 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.198404 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aae4e0f1-0295-4964-8b34-56bdab58a152-combined-ca-bundle\") pod \"placement-db-sync-ql4lt\" (UID: \"aae4e0f1-0295-4964-8b34-56bdab58a152\") " pod="openstack/placement-db-sync-ql4lt" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.198574 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aae4e0f1-0295-4964-8b34-56bdab58a152-scripts\") pod \"placement-db-sync-ql4lt\" (UID: \"aae4e0f1-0295-4964-8b34-56bdab58a152\") " pod="openstack/placement-db-sync-ql4lt" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.198622 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.198759 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.198903 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.198976 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.199134 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.199194 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aae4e0f1-0295-4964-8b34-56bdab58a152-logs\") pod \"placement-db-sync-ql4lt\" (UID: \"aae4e0f1-0295-4964-8b34-56bdab58a152\") " pod="openstack/placement-db-sync-ql4lt" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.199350 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-logs\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.199406 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aae4e0f1-0295-4964-8b34-56bdab58a152-config-data\") pod \"placement-db-sync-ql4lt\" (UID: \"aae4e0f1-0295-4964-8b34-56bdab58a152\") " pod="openstack/placement-db-sync-ql4lt" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.199497 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt9zl\" (UniqueName: \"kubernetes.io/projected/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-kube-api-access-mt9zl\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.200505 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.201829 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-logs\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.202182 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aae4e0f1-0295-4964-8b34-56bdab58a152-logs\") pod \"placement-db-sync-ql4lt\" (UID: \"aae4e0f1-0295-4964-8b34-56bdab58a152\") " pod="openstack/placement-db-sync-ql4lt" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.246277 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.246327 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/50da71ebd66a0437f4c3596841b2bf86ae7bfd6dc1ba487324809d1cf1c5f125/globalmount\"" pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.298363 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aae4e0f1-0295-4964-8b34-56bdab58a152-scripts\") pod \"placement-db-sync-ql4lt\" (UID: \"aae4e0f1-0295-4964-8b34-56bdab58a152\") " pod="openstack/placement-db-sync-ql4lt" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.299685 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-57rtp"] Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.303688 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45021af5-224b-4139-a7b6-06ce097681f4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.306591 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-57rtp\" (UID: \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\") " pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.306673 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4qnv\" (UniqueName: \"kubernetes.io/projected/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-kube-api-access-t4qnv\") pod \"dnsmasq-dns-785d8bcb8c-57rtp\" (UID: \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\") " pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.307033 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/232c4238-3f59-4e05-ac75-8746ab0a0069-log-httpd\") pod \"ceilometer-0\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " pod="openstack/ceilometer-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.307129 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/232c4238-3f59-4e05-ac75-8746ab0a0069-run-httpd\") pod \"ceilometer-0\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " pod="openstack/ceilometer-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.307191 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6zsx\" (UniqueName: \"kubernetes.io/projected/232c4238-3f59-4e05-ac75-8746ab0a0069-kube-api-access-r6zsx\") pod \"ceilometer-0\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " pod="openstack/ceilometer-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.307440 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-config\") pod \"dnsmasq-dns-785d8bcb8c-57rtp\" (UID: \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\") " pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.307519 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/232c4238-3f59-4e05-ac75-8746ab0a0069-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " pod="openstack/ceilometer-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.307671 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-57rtp\" (UID: \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\") " pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.308071 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-57rtp\" (UID: \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\") " pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.308105 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/232c4238-3f59-4e05-ac75-8746ab0a0069-config-data\") pod \"ceilometer-0\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " pod="openstack/ceilometer-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.308151 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-57rtp\" (UID: \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\") " pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.308451 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/232c4238-3f59-4e05-ac75-8746ab0a0069-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " pod="openstack/ceilometer-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.308564 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/232c4238-3f59-4e05-ac75-8746ab0a0069-scripts\") pod \"ceilometer-0\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " pod="openstack/ceilometer-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.309852 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.310164 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aae4e0f1-0295-4964-8b34-56bdab58a152-config-data\") pod \"placement-db-sync-ql4lt\" (UID: \"aae4e0f1-0295-4964-8b34-56bdab58a152\") " pod="openstack/placement-db-sync-ql4lt" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.311489 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aae4e0f1-0295-4964-8b34-56bdab58a152-combined-ca-bundle\") pod \"placement-db-sync-ql4lt\" (UID: \"aae4e0f1-0295-4964-8b34-56bdab58a152\") " pod="openstack/placement-db-sync-ql4lt" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.313475 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.316531 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.318300 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxs6n\" (UniqueName: \"kubernetes.io/projected/45021af5-224b-4139-a7b6-06ce097681f4-kube-api-access-lxs6n\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.318554 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz2wp\" (UniqueName: \"kubernetes.io/projected/aae4e0f1-0295-4964-8b34-56bdab58a152-kube-api-access-tz2wp\") pod \"placement-db-sync-ql4lt\" (UID: \"aae4e0f1-0295-4964-8b34-56bdab58a152\") " pod="openstack/placement-db-sync-ql4lt" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.327938 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt9zl\" (UniqueName: \"kubernetes.io/projected/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-kube-api-access-mt9zl\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.330928 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.372957 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\") pod \"glance-default-external-api-0\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.409139 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\") pod \"glance-default-internal-api-0\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.411555 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-config\") pod \"dnsmasq-dns-785d8bcb8c-57rtp\" (UID: \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\") " pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.411651 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/232c4238-3f59-4e05-ac75-8746ab0a0069-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " pod="openstack/ceilometer-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.411753 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-57rtp\" (UID: \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\") " pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.416053 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-57rtp\" (UID: \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\") " pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.416106 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/232c4238-3f59-4e05-ac75-8746ab0a0069-config-data\") pod \"ceilometer-0\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " pod="openstack/ceilometer-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.416150 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-57rtp\" (UID: \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\") " pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.441764 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-57rtp\" (UID: \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\") " pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.442685 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-config\") pod \"dnsmasq-dns-785d8bcb8c-57rtp\" (UID: \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\") " pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.447164 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/232c4238-3f59-4e05-ac75-8746ab0a0069-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " pod="openstack/ceilometer-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.545447 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-57rtp\" (UID: \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\") " pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.545517 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/232c4238-3f59-4e05-ac75-8746ab0a0069-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " pod="openstack/ceilometer-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.545588 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/232c4238-3f59-4e05-ac75-8746ab0a0069-scripts\") pod \"ceilometer-0\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " pod="openstack/ceilometer-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.545740 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-57rtp\" (UID: \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\") " pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.545819 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4qnv\" (UniqueName: \"kubernetes.io/projected/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-kube-api-access-t4qnv\") pod \"dnsmasq-dns-785d8bcb8c-57rtp\" (UID: \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\") " pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.545961 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/232c4238-3f59-4e05-ac75-8746ab0a0069-log-httpd\") pod \"ceilometer-0\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " pod="openstack/ceilometer-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.546014 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/232c4238-3f59-4e05-ac75-8746ab0a0069-run-httpd\") pod \"ceilometer-0\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " pod="openstack/ceilometer-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.546054 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6zsx\" (UniqueName: \"kubernetes.io/projected/232c4238-3f59-4e05-ac75-8746ab0a0069-kube-api-access-r6zsx\") pod \"ceilometer-0\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " pod="openstack/ceilometer-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.550579 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-57rtp\" (UID: \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\") " pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.554554 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-57rtp\" (UID: \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\") " pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.555738 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/232c4238-3f59-4e05-ac75-8746ab0a0069-log-httpd\") pod \"ceilometer-0\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " pod="openstack/ceilometer-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.555996 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/232c4238-3f59-4e05-ac75-8746ab0a0069-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " pod="openstack/ceilometer-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.556331 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/232c4238-3f59-4e05-ac75-8746ab0a0069-run-httpd\") pod \"ceilometer-0\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " pod="openstack/ceilometer-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.567772 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/232c4238-3f59-4e05-ac75-8746ab0a0069-config-data\") pod \"ceilometer-0\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " pod="openstack/ceilometer-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.582139 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/232c4238-3f59-4e05-ac75-8746ab0a0069-scripts\") pod \"ceilometer-0\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " pod="openstack/ceilometer-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.588550 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4qnv\" (UniqueName: \"kubernetes.io/projected/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-kube-api-access-t4qnv\") pod \"dnsmasq-dns-785d8bcb8c-57rtp\" (UID: \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\") " pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.593471 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6zsx\" (UniqueName: \"kubernetes.io/projected/232c4238-3f59-4e05-ac75-8746ab0a0069-kube-api-access-r6zsx\") pod \"ceilometer-0\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " pod="openstack/ceilometer-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.595272 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.619409 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 09 10:11:16 crc kubenswrapper[4824]: I1209 10:11:16.944335 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ql4lt" Dec 09 10:11:17 crc kubenswrapper[4824]: I1209 10:11:17.004403 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:11:17 crc kubenswrapper[4824]: I1209 10:11:17.011014 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-st484"] Dec 09 10:11:17 crc kubenswrapper[4824]: I1209 10:11:17.040812 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" Dec 09 10:11:17 crc kubenswrapper[4824]: I1209 10:11:17.152489 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6lhrx"] Dec 09 10:11:17 crc kubenswrapper[4824]: I1209 10:11:17.733772 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-vfnsw"] Dec 09 10:11:17 crc kubenswrapper[4824]: I1209 10:11:17.743943 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-mqjtv"] Dec 09 10:11:17 crc kubenswrapper[4824]: I1209 10:11:17.756457 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-jghjq"] Dec 09 10:11:17 crc kubenswrapper[4824]: I1209 10:11:17.767678 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-d4klr"] Dec 09 10:11:17 crc kubenswrapper[4824]: W1209 10:11:17.804758 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod405d0a5e_fd62_4146_bfb9_96ff6c077836.slice/crio-e09a391a1ba56ef305ee8e1a594b4b370dc65c7068801fc2b2a7eba8a2015ce8 WatchSource:0}: Error finding container e09a391a1ba56ef305ee8e1a594b4b370dc65c7068801fc2b2a7eba8a2015ce8: Status 404 returned error can't find the container with id e09a391a1ba56ef305ee8e1a594b4b370dc65c7068801fc2b2a7eba8a2015ce8 Dec 09 10:11:17 crc kubenswrapper[4824]: I1209 10:11:17.926759 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mqjtv" event={"ID":"8eb90797-2763-4ffb-8fa5-32a1c8db5c84","Type":"ContainerStarted","Data":"ae2b0b9451945375a741db97237b099463efb38f3e2a7f706dc2ee68147ed469"} Dec 09 10:11:17 crc kubenswrapper[4824]: I1209 10:11:17.939895 4824 generic.go:334] "Generic (PLEG): container finished" podID="a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021" containerID="93205abb68486a3c737d911924713e30087f7b332ea62b542f4adeffd51f0720" exitCode=0 Dec 09 10:11:17 crc kubenswrapper[4824]: I1209 10:11:17.940196 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-st484" event={"ID":"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021","Type":"ContainerDied","Data":"93205abb68486a3c737d911924713e30087f7b332ea62b542f4adeffd51f0720"} Dec 09 10:11:17 crc kubenswrapper[4824]: I1209 10:11:17.940233 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-st484" event={"ID":"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021","Type":"ContainerStarted","Data":"2fb63b7bf07700aa1c1113df35cd0ccc79726c61a491c9c6da01db8f9d3df32f"} Dec 09 10:11:17 crc kubenswrapper[4824]: I1209 10:11:17.943187 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-jghjq" event={"ID":"dbceafb7-272a-45fc-9b56-df631986f90a","Type":"ContainerStarted","Data":"1c49fa2af8410be9172d0e097fdd5701dbea33fbc41b5f7c18ae4593604a9f73"} Dec 09 10:11:17 crc kubenswrapper[4824]: I1209 10:11:17.983291 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6lhrx" event={"ID":"c7e2a035-7568-4455-a3bb-329dcf3276ca","Type":"ContainerStarted","Data":"aa6386d2497fdcabbe806274e4229a9e73ef6471b24447710b88924b91553266"} Dec 09 10:11:17 crc kubenswrapper[4824]: I1209 10:11:17.983691 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6lhrx" event={"ID":"c7e2a035-7568-4455-a3bb-329dcf3276ca","Type":"ContainerStarted","Data":"aa8abf351d6d714313c098612fd6fe9873e75ba590ea98dc826abdd71cb9a285"} Dec 09 10:11:18 crc kubenswrapper[4824]: I1209 10:11:18.019956 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-vfnsw" event={"ID":"405d0a5e-fd62-4146-bfb9-96ff6c077836","Type":"ContainerStarted","Data":"e09a391a1ba56ef305ee8e1a594b4b370dc65c7068801fc2b2a7eba8a2015ce8"} Dec 09 10:11:18 crc kubenswrapper[4824]: I1209 10:11:18.047896 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-6lhrx" podStartSLOduration=5.047864138 podStartE2EDuration="5.047864138s" podCreationTimestamp="2025-12-09 10:11:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:11:18.005433608 +0000 UTC m=+1434.339938275" watchObservedRunningTime="2025-12-09 10:11:18.047864138 +0000 UTC m=+1434.382368805" Dec 09 10:11:18 crc kubenswrapper[4824]: I1209 10:11:18.096833 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-ql4lt"] Dec 09 10:11:18 crc kubenswrapper[4824]: I1209 10:11:18.129083 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 10:11:18 crc kubenswrapper[4824]: I1209 10:11:18.225988 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-57rtp"] Dec 09 10:11:18 crc kubenswrapper[4824]: I1209 10:11:18.321665 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:11:18 crc kubenswrapper[4824]: I1209 10:11:18.826480 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 10:11:18 crc kubenswrapper[4824]: I1209 10:11:18.848304 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-st484" Dec 09 10:11:18 crc kubenswrapper[4824]: I1209 10:11:18.861004 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-ovsdbserver-nb\") pod \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\" (UID: \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\") " Dec 09 10:11:18 crc kubenswrapper[4824]: I1209 10:11:18.861210 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9tfhn\" (UniqueName: \"kubernetes.io/projected/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-kube-api-access-9tfhn\") pod \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\" (UID: \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\") " Dec 09 10:11:18 crc kubenswrapper[4824]: I1209 10:11:18.861300 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-ovsdbserver-sb\") pod \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\" (UID: \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\") " Dec 09 10:11:18 crc kubenswrapper[4824]: I1209 10:11:18.861556 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-dns-svc\") pod \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\" (UID: \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\") " Dec 09 10:11:18 crc kubenswrapper[4824]: I1209 10:11:18.861636 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-dns-swift-storage-0\") pod \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\" (UID: \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\") " Dec 09 10:11:18 crc kubenswrapper[4824]: I1209 10:11:18.861748 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-config\") pod \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\" (UID: \"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021\") " Dec 09 10:11:18 crc kubenswrapper[4824]: I1209 10:11:18.894005 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-kube-api-access-9tfhn" (OuterVolumeSpecName: "kube-api-access-9tfhn") pod "a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021" (UID: "a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021"). InnerVolumeSpecName "kube-api-access-9tfhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:11:18 crc kubenswrapper[4824]: I1209 10:11:18.946968 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021" (UID: "a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:11:18 crc kubenswrapper[4824]: I1209 10:11:18.970942 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-config" (OuterVolumeSpecName: "config") pod "a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021" (UID: "a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:11:18 crc kubenswrapper[4824]: I1209 10:11:18.970993 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:18 crc kubenswrapper[4824]: I1209 10:11:18.971019 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9tfhn\" (UniqueName: \"kubernetes.io/projected/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-kube-api-access-9tfhn\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:18 crc kubenswrapper[4824]: I1209 10:11:18.971306 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021" (UID: "a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:11:19 crc kubenswrapper[4824]: I1209 10:11:19.033855 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021" (UID: "a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:11:19 crc kubenswrapper[4824]: I1209 10:11:19.033981 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021" (UID: "a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:11:19 crc kubenswrapper[4824]: I1209 10:11:19.073507 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:19 crc kubenswrapper[4824]: I1209 10:11:19.073557 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:19 crc kubenswrapper[4824]: I1209 10:11:19.073575 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:19 crc kubenswrapper[4824]: I1209 10:11:19.073586 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:19 crc kubenswrapper[4824]: I1209 10:11:19.153256 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"45021af5-224b-4139-a7b6-06ce097681f4","Type":"ContainerStarted","Data":"664ee94ee7b6f41357453cd8dc3a6ee6bad8309748fcd3b63c73629da15ed3b2"} Dec 09 10:11:19 crc kubenswrapper[4824]: I1209 10:11:19.156917 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" event={"ID":"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31","Type":"ContainerStarted","Data":"61b4effa6402fb58c433e87b2da3cb8f5ec47876048837b2193ab8e4b7737e24"} Dec 09 10:11:19 crc kubenswrapper[4824]: I1209 10:11:19.174800 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mqjtv" event={"ID":"8eb90797-2763-4ffb-8fa5-32a1c8db5c84","Type":"ContainerStarted","Data":"481598142ccbbb37a8c1d4b3c9c41b36e6797b272a9cd42c11a24be6d054d4d6"} Dec 09 10:11:19 crc kubenswrapper[4824]: I1209 10:11:19.187957 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-st484" event={"ID":"a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021","Type":"ContainerDied","Data":"2fb63b7bf07700aa1c1113df35cd0ccc79726c61a491c9c6da01db8f9d3df32f"} Dec 09 10:11:19 crc kubenswrapper[4824]: I1209 10:11:19.188025 4824 scope.go:117] "RemoveContainer" containerID="93205abb68486a3c737d911924713e30087f7b332ea62b542f4adeffd51f0720" Dec 09 10:11:19 crc kubenswrapper[4824]: I1209 10:11:19.188239 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-st484" Dec 09 10:11:19 crc kubenswrapper[4824]: I1209 10:11:19.205400 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5d1dbb17-3c4e-4801-bf8b-a29806de5c54","Type":"ContainerStarted","Data":"9fcbe41babe98705696ca3ab9f99f1056c113053f1b981b3e54fcba2b6b41158"} Dec 09 10:11:19 crc kubenswrapper[4824]: I1209 10:11:19.214619 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-mqjtv" podStartSLOduration=5.214587754 podStartE2EDuration="5.214587754s" podCreationTimestamp="2025-12-09 10:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:11:19.197301354 +0000 UTC m=+1435.531806031" watchObservedRunningTime="2025-12-09 10:11:19.214587754 +0000 UTC m=+1435.549092421" Dec 09 10:11:19 crc kubenswrapper[4824]: I1209 10:11:19.217948 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ql4lt" event={"ID":"aae4e0f1-0295-4964-8b34-56bdab58a152","Type":"ContainerStarted","Data":"445f3f2271e4d78cb53e9d0ce785081f27feb5a77762defdfb2a15209f6d44a3"} Dec 09 10:11:19 crc kubenswrapper[4824]: I1209 10:11:19.221962 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-d4klr" event={"ID":"d3d076df-d1e1-473e-9584-79b30cb6d023","Type":"ContainerStarted","Data":"8857606b103f540643d96ffaa2f25674c464039ebffbf499c32645926ab0d0f8"} Dec 09 10:11:19 crc kubenswrapper[4824]: I1209 10:11:19.226310 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"232c4238-3f59-4e05-ac75-8746ab0a0069","Type":"ContainerStarted","Data":"ed62ce1026e7eb19f4d8be85136efcd90bca262ff995fe62cb129a67e7e968de"} Dec 09 10:11:19 crc kubenswrapper[4824]: I1209 10:11:19.309556 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-st484"] Dec 09 10:11:19 crc kubenswrapper[4824]: I1209 10:11:19.324322 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-st484"] Dec 09 10:11:19 crc kubenswrapper[4824]: I1209 10:11:19.951546 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021" path="/var/lib/kubelet/pods/a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021/volumes" Dec 09 10:11:20 crc kubenswrapper[4824]: I1209 10:11:20.147016 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 10:11:20 crc kubenswrapper[4824]: I1209 10:11:20.450966 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 10:11:20 crc kubenswrapper[4824]: I1209 10:11:20.614043 4824 generic.go:334] "Generic (PLEG): container finished" podID="ff1309c3-9703-4b9c-8ec7-e9fdeee87b31" containerID="cc96b5d9ac0d22837e051873be369805cd0878b0bb2394919a4662ab448cb53b" exitCode=0 Dec 09 10:11:20 crc kubenswrapper[4824]: I1209 10:11:20.615866 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" event={"ID":"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31","Type":"ContainerDied","Data":"cc96b5d9ac0d22837e051873be369805cd0878b0bb2394919a4662ab448cb53b"} Dec 09 10:11:20 crc kubenswrapper[4824]: I1209 10:11:20.775712 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:11:21 crc kubenswrapper[4824]: I1209 10:11:21.868211 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"45021af5-224b-4139-a7b6-06ce097681f4","Type":"ContainerStarted","Data":"270a015831ebac9b4ca1a30733ca8ef7d77be57e1e8a715249e2282449121282"} Dec 09 10:11:21 crc kubenswrapper[4824]: I1209 10:11:21.872537 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" event={"ID":"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31","Type":"ContainerStarted","Data":"75dd9380b91d13bbd28f9a9a2502d7f35cef1d2c681580f2b49e40516fe958ac"} Dec 09 10:11:21 crc kubenswrapper[4824]: I1209 10:11:21.896578 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5d1dbb17-3c4e-4801-bf8b-a29806de5c54","Type":"ContainerStarted","Data":"a527c7cbd114a619e610d41cb2f842145ef7965ca4c2a8aaacd892e7e62ec41d"} Dec 09 10:11:22 crc kubenswrapper[4824]: I1209 10:11:22.966397 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5d1dbb17-3c4e-4801-bf8b-a29806de5c54","Type":"ContainerStarted","Data":"7bd036ca098ef02bb8a3b8c091bfc52a8bfa99738bbfa4f2fd0b153e29957144"} Dec 09 10:11:22 crc kubenswrapper[4824]: I1209 10:11:22.966961 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" Dec 09 10:11:22 crc kubenswrapper[4824]: I1209 10:11:22.967396 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5d1dbb17-3c4e-4801-bf8b-a29806de5c54" containerName="glance-httpd" containerID="cri-o://7bd036ca098ef02bb8a3b8c091bfc52a8bfa99738bbfa4f2fd0b153e29957144" gracePeriod=30 Dec 09 10:11:22 crc kubenswrapper[4824]: I1209 10:11:22.967450 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5d1dbb17-3c4e-4801-bf8b-a29806de5c54" containerName="glance-log" containerID="cri-o://a527c7cbd114a619e610d41cb2f842145ef7965ca4c2a8aaacd892e7e62ec41d" gracePeriod=30 Dec 09 10:11:23 crc kubenswrapper[4824]: I1209 10:11:23.012667 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" podStartSLOduration=8.012645074 podStartE2EDuration="8.012645074s" podCreationTimestamp="2025-12-09 10:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:11:23.000010372 +0000 UTC m=+1439.334515039" watchObservedRunningTime="2025-12-09 10:11:23.012645074 +0000 UTC m=+1439.347149741" Dec 09 10:11:23 crc kubenswrapper[4824]: I1209 10:11:23.049645 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=9.049614831 podStartE2EDuration="9.049614831s" podCreationTimestamp="2025-12-09 10:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:11:23.034136207 +0000 UTC m=+1439.368640884" watchObservedRunningTime="2025-12-09 10:11:23.049614831 +0000 UTC m=+1439.384119498" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.002248 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.003042 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="45021af5-224b-4139-a7b6-06ce097681f4" containerName="glance-log" containerID="cri-o://270a015831ebac9b4ca1a30733ca8ef7d77be57e1e8a715249e2282449121282" gracePeriod=30 Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.003082 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="45021af5-224b-4139-a7b6-06ce097681f4" containerName="glance-httpd" containerID="cri-o://4fa091e2a1af246edfc52ed281184b45738b224d35cb4a55c08225c1268c7d1c" gracePeriod=30 Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.009562 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"45021af5-224b-4139-a7b6-06ce097681f4","Type":"ContainerStarted","Data":"4fa091e2a1af246edfc52ed281184b45738b224d35cb4a55c08225c1268c7d1c"} Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.014601 4824 generic.go:334] "Generic (PLEG): container finished" podID="5d1dbb17-3c4e-4801-bf8b-a29806de5c54" containerID="7bd036ca098ef02bb8a3b8c091bfc52a8bfa99738bbfa4f2fd0b153e29957144" exitCode=143 Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.014652 4824 generic.go:334] "Generic (PLEG): container finished" podID="5d1dbb17-3c4e-4801-bf8b-a29806de5c54" containerID="a527c7cbd114a619e610d41cb2f842145ef7965ca4c2a8aaacd892e7e62ec41d" exitCode=143 Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.015270 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5d1dbb17-3c4e-4801-bf8b-a29806de5c54","Type":"ContainerDied","Data":"7bd036ca098ef02bb8a3b8c091bfc52a8bfa99738bbfa4f2fd0b153e29957144"} Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.015360 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5d1dbb17-3c4e-4801-bf8b-a29806de5c54","Type":"ContainerDied","Data":"a527c7cbd114a619e610d41cb2f842145ef7965ca4c2a8aaacd892e7e62ec41d"} Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.015375 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5d1dbb17-3c4e-4801-bf8b-a29806de5c54","Type":"ContainerDied","Data":"9fcbe41babe98705696ca3ab9f99f1056c113053f1b981b3e54fcba2b6b41158"} Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.015397 4824 scope.go:117] "RemoveContainer" containerID="7bd036ca098ef02bb8a3b8c091bfc52a8bfa99738bbfa4f2fd0b153e29957144" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.063568 4824 scope.go:117] "RemoveContainer" containerID="a527c7cbd114a619e610d41cb2f842145ef7965ca4c2a8aaacd892e7e62ec41d" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.079640 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-scripts\") pod \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.079836 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-httpd-run\") pod \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.079932 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-logs\") pod \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.079968 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-combined-ca-bundle\") pod \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.079996 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-config-data\") pod \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.080053 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mt9zl\" (UniqueName: \"kubernetes.io/projected/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-kube-api-access-mt9zl\") pod \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.080144 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-internal-tls-certs\") pod \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.080226 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5d1dbb17-3c4e-4801-bf8b-a29806de5c54" (UID: "5d1dbb17-3c4e-4801-bf8b-a29806de5c54"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.080310 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\") pod \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\" (UID: \"5d1dbb17-3c4e-4801-bf8b-a29806de5c54\") " Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.080700 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-logs" (OuterVolumeSpecName: "logs") pod "5d1dbb17-3c4e-4801-bf8b-a29806de5c54" (UID: "5d1dbb17-3c4e-4801-bf8b-a29806de5c54"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.081820 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-logs\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.081849 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.091197 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-scripts" (OuterVolumeSpecName: "scripts") pod "5d1dbb17-3c4e-4801-bf8b-a29806de5c54" (UID: "5d1dbb17-3c4e-4801-bf8b-a29806de5c54"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.103039 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-kube-api-access-mt9zl" (OuterVolumeSpecName: "kube-api-access-mt9zl") pod "5d1dbb17-3c4e-4801-bf8b-a29806de5c54" (UID: "5d1dbb17-3c4e-4801-bf8b-a29806de5c54"). InnerVolumeSpecName "kube-api-access-mt9zl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.125712 4824 scope.go:117] "RemoveContainer" containerID="7bd036ca098ef02bb8a3b8c091bfc52a8bfa99738bbfa4f2fd0b153e29957144" Dec 09 10:11:24 crc kubenswrapper[4824]: E1209 10:11:24.128758 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bd036ca098ef02bb8a3b8c091bfc52a8bfa99738bbfa4f2fd0b153e29957144\": container with ID starting with 7bd036ca098ef02bb8a3b8c091bfc52a8bfa99738bbfa4f2fd0b153e29957144 not found: ID does not exist" containerID="7bd036ca098ef02bb8a3b8c091bfc52a8bfa99738bbfa4f2fd0b153e29957144" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.128823 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bd036ca098ef02bb8a3b8c091bfc52a8bfa99738bbfa4f2fd0b153e29957144"} err="failed to get container status \"7bd036ca098ef02bb8a3b8c091bfc52a8bfa99738bbfa4f2fd0b153e29957144\": rpc error: code = NotFound desc = could not find container \"7bd036ca098ef02bb8a3b8c091bfc52a8bfa99738bbfa4f2fd0b153e29957144\": container with ID starting with 7bd036ca098ef02bb8a3b8c091bfc52a8bfa99738bbfa4f2fd0b153e29957144 not found: ID does not exist" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.128850 4824 scope.go:117] "RemoveContainer" containerID="a527c7cbd114a619e610d41cb2f842145ef7965ca4c2a8aaacd892e7e62ec41d" Dec 09 10:11:24 crc kubenswrapper[4824]: E1209 10:11:24.129199 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a527c7cbd114a619e610d41cb2f842145ef7965ca4c2a8aaacd892e7e62ec41d\": container with ID starting with a527c7cbd114a619e610d41cb2f842145ef7965ca4c2a8aaacd892e7e62ec41d not found: ID does not exist" containerID="a527c7cbd114a619e610d41cb2f842145ef7965ca4c2a8aaacd892e7e62ec41d" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.129268 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a527c7cbd114a619e610d41cb2f842145ef7965ca4c2a8aaacd892e7e62ec41d"} err="failed to get container status \"a527c7cbd114a619e610d41cb2f842145ef7965ca4c2a8aaacd892e7e62ec41d\": rpc error: code = NotFound desc = could not find container \"a527c7cbd114a619e610d41cb2f842145ef7965ca4c2a8aaacd892e7e62ec41d\": container with ID starting with a527c7cbd114a619e610d41cb2f842145ef7965ca4c2a8aaacd892e7e62ec41d not found: ID does not exist" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.129503 4824 scope.go:117] "RemoveContainer" containerID="7bd036ca098ef02bb8a3b8c091bfc52a8bfa99738bbfa4f2fd0b153e29957144" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.130515 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bd036ca098ef02bb8a3b8c091bfc52a8bfa99738bbfa4f2fd0b153e29957144"} err="failed to get container status \"7bd036ca098ef02bb8a3b8c091bfc52a8bfa99738bbfa4f2fd0b153e29957144\": rpc error: code = NotFound desc = could not find container \"7bd036ca098ef02bb8a3b8c091bfc52a8bfa99738bbfa4f2fd0b153e29957144\": container with ID starting with 7bd036ca098ef02bb8a3b8c091bfc52a8bfa99738bbfa4f2fd0b153e29957144 not found: ID does not exist" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.130604 4824 scope.go:117] "RemoveContainer" containerID="a527c7cbd114a619e610d41cb2f842145ef7965ca4c2a8aaacd892e7e62ec41d" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.131399 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a527c7cbd114a619e610d41cb2f842145ef7965ca4c2a8aaacd892e7e62ec41d"} err="failed to get container status \"a527c7cbd114a619e610d41cb2f842145ef7965ca4c2a8aaacd892e7e62ec41d\": rpc error: code = NotFound desc = could not find container \"a527c7cbd114a619e610d41cb2f842145ef7965ca4c2a8aaacd892e7e62ec41d\": container with ID starting with a527c7cbd114a619e610d41cb2f842145ef7965ca4c2a8aaacd892e7e62ec41d not found: ID does not exist" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.152546 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d1dbb17-3c4e-4801-bf8b-a29806de5c54" (UID: "5d1dbb17-3c4e-4801-bf8b-a29806de5c54"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.166856 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69" (OuterVolumeSpecName: "glance") pod "5d1dbb17-3c4e-4801-bf8b-a29806de5c54" (UID: "5d1dbb17-3c4e-4801-bf8b-a29806de5c54"). InnerVolumeSpecName "pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.180584 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=10.180557248 podStartE2EDuration="10.180557248s" podCreationTimestamp="2025-12-09 10:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:11:24.171015884 +0000 UTC m=+1440.505520551" watchObservedRunningTime="2025-12-09 10:11:24.180557248 +0000 UTC m=+1440.515061915" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.184676 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.184719 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mt9zl\" (UniqueName: \"kubernetes.io/projected/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-kube-api-access-mt9zl\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.184771 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\") on node \"crc\" " Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.184804 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.229021 4824 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.229302 4824 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69") on node "crc" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.231540 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5d1dbb17-3c4e-4801-bf8b-a29806de5c54" (UID: "5d1dbb17-3c4e-4801-bf8b-a29806de5c54"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.292744 4824 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.299025 4824 reconciler_common.go:293] "Volume detached for volume \"pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.308096 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-config-data" (OuterVolumeSpecName: "config-data") pod "5d1dbb17-3c4e-4801-bf8b-a29806de5c54" (UID: "5d1dbb17-3c4e-4801-bf8b-a29806de5c54"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:11:24 crc kubenswrapper[4824]: I1209 10:11:24.401198 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d1dbb17-3c4e-4801-bf8b-a29806de5c54-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.061626 4824 generic.go:334] "Generic (PLEG): container finished" podID="c7e2a035-7568-4455-a3bb-329dcf3276ca" containerID="aa6386d2497fdcabbe806274e4229a9e73ef6471b24447710b88924b91553266" exitCode=0 Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.061757 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6lhrx" event={"ID":"c7e2a035-7568-4455-a3bb-329dcf3276ca","Type":"ContainerDied","Data":"aa6386d2497fdcabbe806274e4229a9e73ef6471b24447710b88924b91553266"} Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.065264 4824 generic.go:334] "Generic (PLEG): container finished" podID="45021af5-224b-4139-a7b6-06ce097681f4" containerID="4fa091e2a1af246edfc52ed281184b45738b224d35cb4a55c08225c1268c7d1c" exitCode=0 Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.065297 4824 generic.go:334] "Generic (PLEG): container finished" podID="45021af5-224b-4139-a7b6-06ce097681f4" containerID="270a015831ebac9b4ca1a30733ca8ef7d77be57e1e8a715249e2282449121282" exitCode=143 Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.065351 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"45021af5-224b-4139-a7b6-06ce097681f4","Type":"ContainerDied","Data":"4fa091e2a1af246edfc52ed281184b45738b224d35cb4a55c08225c1268c7d1c"} Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.065388 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"45021af5-224b-4139-a7b6-06ce097681f4","Type":"ContainerDied","Data":"270a015831ebac9b4ca1a30733ca8ef7d77be57e1e8a715249e2282449121282"} Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.068096 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.138933 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.161961 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.184538 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.224296 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 10:11:25 crc kubenswrapper[4824]: E1209 10:11:25.225087 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d1dbb17-3c4e-4801-bf8b-a29806de5c54" containerName="glance-log" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.225122 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d1dbb17-3c4e-4801-bf8b-a29806de5c54" containerName="glance-log" Dec 09 10:11:25 crc kubenswrapper[4824]: E1209 10:11:25.225146 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021" containerName="init" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.225156 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021" containerName="init" Dec 09 10:11:25 crc kubenswrapper[4824]: E1209 10:11:25.225172 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45021af5-224b-4139-a7b6-06ce097681f4" containerName="glance-httpd" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.225181 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="45021af5-224b-4139-a7b6-06ce097681f4" containerName="glance-httpd" Dec 09 10:11:25 crc kubenswrapper[4824]: E1209 10:11:25.225207 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d1dbb17-3c4e-4801-bf8b-a29806de5c54" containerName="glance-httpd" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.225216 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d1dbb17-3c4e-4801-bf8b-a29806de5c54" containerName="glance-httpd" Dec 09 10:11:25 crc kubenswrapper[4824]: E1209 10:11:25.225258 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45021af5-224b-4139-a7b6-06ce097681f4" containerName="glance-log" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.225266 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="45021af5-224b-4139-a7b6-06ce097681f4" containerName="glance-log" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.225572 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d1dbb17-3c4e-4801-bf8b-a29806de5c54" containerName="glance-httpd" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.225598 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7d3aca5-2d6d-4a1c-b3a8-e322b9df2021" containerName="init" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.225611 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="45021af5-224b-4139-a7b6-06ce097681f4" containerName="glance-httpd" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.225623 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d1dbb17-3c4e-4801-bf8b-a29806de5c54" containerName="glance-log" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.225659 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="45021af5-224b-4139-a7b6-06ce097681f4" containerName="glance-log" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.227527 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.231820 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.232700 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.279855 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.326492 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxs6n\" (UniqueName: \"kubernetes.io/projected/45021af5-224b-4139-a7b6-06ce097681f4-kube-api-access-lxs6n\") pod \"45021af5-224b-4139-a7b6-06ce097681f4\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.326546 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45021af5-224b-4139-a7b6-06ce097681f4-public-tls-certs\") pod \"45021af5-224b-4139-a7b6-06ce097681f4\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.326643 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45021af5-224b-4139-a7b6-06ce097681f4-logs\") pod \"45021af5-224b-4139-a7b6-06ce097681f4\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.326734 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/45021af5-224b-4139-a7b6-06ce097681f4-httpd-run\") pod \"45021af5-224b-4139-a7b6-06ce097681f4\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.326931 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45021af5-224b-4139-a7b6-06ce097681f4-scripts\") pod \"45021af5-224b-4139-a7b6-06ce097681f4\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.327080 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\") pod \"45021af5-224b-4139-a7b6-06ce097681f4\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.327129 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45021af5-224b-4139-a7b6-06ce097681f4-config-data\") pod \"45021af5-224b-4139-a7b6-06ce097681f4\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.327152 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45021af5-224b-4139-a7b6-06ce097681f4-combined-ca-bundle\") pod \"45021af5-224b-4139-a7b6-06ce097681f4\" (UID: \"45021af5-224b-4139-a7b6-06ce097681f4\") " Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.328517 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45021af5-224b-4139-a7b6-06ce097681f4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "45021af5-224b-4139-a7b6-06ce097681f4" (UID: "45021af5-224b-4139-a7b6-06ce097681f4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.331188 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45021af5-224b-4139-a7b6-06ce097681f4-logs" (OuterVolumeSpecName: "logs") pod "45021af5-224b-4139-a7b6-06ce097681f4" (UID: "45021af5-224b-4139-a7b6-06ce097681f4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.351239 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45021af5-224b-4139-a7b6-06ce097681f4-scripts" (OuterVolumeSpecName: "scripts") pod "45021af5-224b-4139-a7b6-06ce097681f4" (UID: "45021af5-224b-4139-a7b6-06ce097681f4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.355626 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd" (OuterVolumeSpecName: "glance") pod "45021af5-224b-4139-a7b6-06ce097681f4" (UID: "45021af5-224b-4139-a7b6-06ce097681f4"). InnerVolumeSpecName "pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.367753 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45021af5-224b-4139-a7b6-06ce097681f4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "45021af5-224b-4139-a7b6-06ce097681f4" (UID: "45021af5-224b-4139-a7b6-06ce097681f4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.371472 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45021af5-224b-4139-a7b6-06ce097681f4-kube-api-access-lxs6n" (OuterVolumeSpecName: "kube-api-access-lxs6n") pod "45021af5-224b-4139-a7b6-06ce097681f4" (UID: "45021af5-224b-4139-a7b6-06ce097681f4"). InnerVolumeSpecName "kube-api-access-lxs6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.429401 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d974a3d8-22a8-4301-bb5f-28011189b4dc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.429720 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d974a3d8-22a8-4301-bb5f-28011189b4dc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.429976 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d974a3d8-22a8-4301-bb5f-28011189b4dc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.430115 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.430225 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d974a3d8-22a8-4301-bb5f-28011189b4dc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.430382 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d974a3d8-22a8-4301-bb5f-28011189b4dc-logs\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.430564 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8lf2\" (UniqueName: \"kubernetes.io/projected/d974a3d8-22a8-4301-bb5f-28011189b4dc-kube-api-access-g8lf2\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.430642 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d974a3d8-22a8-4301-bb5f-28011189b4dc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.431175 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45021af5-224b-4139-a7b6-06ce097681f4-logs\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.431205 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/45021af5-224b-4139-a7b6-06ce097681f4-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.431219 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45021af5-224b-4139-a7b6-06ce097681f4-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.431251 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\") on node \"crc\" " Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.431268 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45021af5-224b-4139-a7b6-06ce097681f4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.431285 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxs6n\" (UniqueName: \"kubernetes.io/projected/45021af5-224b-4139-a7b6-06ce097681f4-kube-api-access-lxs6n\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.431708 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45021af5-224b-4139-a7b6-06ce097681f4-config-data" (OuterVolumeSpecName: "config-data") pod "45021af5-224b-4139-a7b6-06ce097681f4" (UID: "45021af5-224b-4139-a7b6-06ce097681f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.442665 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45021af5-224b-4139-a7b6-06ce097681f4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "45021af5-224b-4139-a7b6-06ce097681f4" (UID: "45021af5-224b-4139-a7b6-06ce097681f4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.493565 4824 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.493750 4824 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd") on node "crc" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.534218 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d974a3d8-22a8-4301-bb5f-28011189b4dc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.534730 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d974a3d8-22a8-4301-bb5f-28011189b4dc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.534758 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d974a3d8-22a8-4301-bb5f-28011189b4dc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.534833 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.534903 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d974a3d8-22a8-4301-bb5f-28011189b4dc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.534980 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d974a3d8-22a8-4301-bb5f-28011189b4dc-logs\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.535124 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8lf2\" (UniqueName: \"kubernetes.io/projected/d974a3d8-22a8-4301-bb5f-28011189b4dc-kube-api-access-g8lf2\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.535180 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d974a3d8-22a8-4301-bb5f-28011189b4dc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.535359 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d974a3d8-22a8-4301-bb5f-28011189b4dc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.535527 4824 reconciler_common.go:293] "Volume detached for volume \"pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.535547 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45021af5-224b-4139-a7b6-06ce097681f4-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.535562 4824 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45021af5-224b-4139-a7b6-06ce097681f4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.537674 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d974a3d8-22a8-4301-bb5f-28011189b4dc-logs\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.539096 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.539134 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/50da71ebd66a0437f4c3596841b2bf86ae7bfd6dc1ba487324809d1cf1c5f125/globalmount\"" pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.540275 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d974a3d8-22a8-4301-bb5f-28011189b4dc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.540375 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d974a3d8-22a8-4301-bb5f-28011189b4dc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.550585 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d974a3d8-22a8-4301-bb5f-28011189b4dc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.557524 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d974a3d8-22a8-4301-bb5f-28011189b4dc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.561109 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8lf2\" (UniqueName: \"kubernetes.io/projected/d974a3d8-22a8-4301-bb5f-28011189b4dc-kube-api-access-g8lf2\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.609986 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\") pod \"glance-default-internal-api-0\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.858685 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 09 10:11:25 crc kubenswrapper[4824]: I1209 10:11:25.931496 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d1dbb17-3c4e-4801-bf8b-a29806de5c54" path="/var/lib/kubelet/pods/5d1dbb17-3c4e-4801-bf8b-a29806de5c54/volumes" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.099568 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"45021af5-224b-4139-a7b6-06ce097681f4","Type":"ContainerDied","Data":"664ee94ee7b6f41357453cd8dc3a6ee6bad8309748fcd3b63c73629da15ed3b2"} Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.099647 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.099723 4824 scope.go:117] "RemoveContainer" containerID="4fa091e2a1af246edfc52ed281184b45738b224d35cb4a55c08225c1268c7d1c" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.138257 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.156853 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.189270 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.191952 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.195519 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.195734 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.222325 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.368909 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/988197de-9531-4cae-92d0-ec7f836a8c73-logs\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.369354 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/988197de-9531-4cae-92d0-ec7f836a8c73-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.369388 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/988197de-9531-4cae-92d0-ec7f836a8c73-scripts\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.369412 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvlqq\" (UniqueName: \"kubernetes.io/projected/988197de-9531-4cae-92d0-ec7f836a8c73-kube-api-access-bvlqq\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.369722 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.369799 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/988197de-9531-4cae-92d0-ec7f836a8c73-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.369941 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/988197de-9531-4cae-92d0-ec7f836a8c73-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.369973 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/988197de-9531-4cae-92d0-ec7f836a8c73-config-data\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.578352 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/988197de-9531-4cae-92d0-ec7f836a8c73-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.578408 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/988197de-9531-4cae-92d0-ec7f836a8c73-config-data\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.578513 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/988197de-9531-4cae-92d0-ec7f836a8c73-logs\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.578876 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/988197de-9531-4cae-92d0-ec7f836a8c73-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.578912 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvlqq\" (UniqueName: \"kubernetes.io/projected/988197de-9531-4cae-92d0-ec7f836a8c73-kube-api-access-bvlqq\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.578936 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/988197de-9531-4cae-92d0-ec7f836a8c73-scripts\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.579203 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.581425 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/988197de-9531-4cae-92d0-ec7f836a8c73-logs\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.585610 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/988197de-9531-4cae-92d0-ec7f836a8c73-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.586731 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/988197de-9531-4cae-92d0-ec7f836a8c73-config-data\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.596535 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/988197de-9531-4cae-92d0-ec7f836a8c73-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.597955 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/988197de-9531-4cae-92d0-ec7f836a8c73-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.603131 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/988197de-9531-4cae-92d0-ec7f836a8c73-scripts\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.604283 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/988197de-9531-4cae-92d0-ec7f836a8c73-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.608351 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.608401 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/afe41c6e88a0ab9650f976db82a109be2feb044bddc024f73824c502bea72164/globalmount\"" pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.628061 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvlqq\" (UniqueName: \"kubernetes.io/projected/988197de-9531-4cae-92d0-ec7f836a8c73-kube-api-access-bvlqq\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.670639 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\") pod \"glance-default-external-api-0\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " pod="openstack/glance-default-external-api-0" Dec 09 10:11:26 crc kubenswrapper[4824]: I1209 10:11:26.835850 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 09 10:11:27 crc kubenswrapper[4824]: I1209 10:11:27.043954 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" Dec 09 10:11:27 crc kubenswrapper[4824]: I1209 10:11:27.137546 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-fgj5w"] Dec 09 10:11:27 crc kubenswrapper[4824]: I1209 10:11:27.137839 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" podUID="6a45d593-c64c-471e-b0fa-171446548a4d" containerName="dnsmasq-dns" containerID="cri-o://085feec377c96e6eba63a801f66855826edb5591113a6501dd0576d7f8ea05d0" gracePeriod=10 Dec 09 10:11:27 crc kubenswrapper[4824]: I1209 10:11:27.935128 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45021af5-224b-4139-a7b6-06ce097681f4" path="/var/lib/kubelet/pods/45021af5-224b-4139-a7b6-06ce097681f4/volumes" Dec 09 10:11:28 crc kubenswrapper[4824]: I1209 10:11:28.155653 4824 generic.go:334] "Generic (PLEG): container finished" podID="6a45d593-c64c-471e-b0fa-171446548a4d" containerID="085feec377c96e6eba63a801f66855826edb5591113a6501dd0576d7f8ea05d0" exitCode=0 Dec 09 10:11:28 crc kubenswrapper[4824]: I1209 10:11:28.155701 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" event={"ID":"6a45d593-c64c-471e-b0fa-171446548a4d","Type":"ContainerDied","Data":"085feec377c96e6eba63a801f66855826edb5591113a6501dd0576d7f8ea05d0"} Dec 09 10:11:29 crc kubenswrapper[4824]: I1209 10:11:29.519133 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" podUID="6a45d593-c64c-471e-b0fa-171446548a4d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.168:5353: connect: connection refused" Dec 09 10:11:32 crc kubenswrapper[4824]: I1209 10:11:32.030113 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6lhrx" Dec 09 10:11:32 crc kubenswrapper[4824]: I1209 10:11:32.176896 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwhzk\" (UniqueName: \"kubernetes.io/projected/c7e2a035-7568-4455-a3bb-329dcf3276ca-kube-api-access-kwhzk\") pod \"c7e2a035-7568-4455-a3bb-329dcf3276ca\" (UID: \"c7e2a035-7568-4455-a3bb-329dcf3276ca\") " Dec 09 10:11:32 crc kubenswrapper[4824]: I1209 10:11:32.176977 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-config-data\") pod \"c7e2a035-7568-4455-a3bb-329dcf3276ca\" (UID: \"c7e2a035-7568-4455-a3bb-329dcf3276ca\") " Dec 09 10:11:32 crc kubenswrapper[4824]: I1209 10:11:32.177153 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-fernet-keys\") pod \"c7e2a035-7568-4455-a3bb-329dcf3276ca\" (UID: \"c7e2a035-7568-4455-a3bb-329dcf3276ca\") " Dec 09 10:11:32 crc kubenswrapper[4824]: I1209 10:11:32.177241 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-combined-ca-bundle\") pod \"c7e2a035-7568-4455-a3bb-329dcf3276ca\" (UID: \"c7e2a035-7568-4455-a3bb-329dcf3276ca\") " Dec 09 10:11:32 crc kubenswrapper[4824]: I1209 10:11:32.177277 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-credential-keys\") pod \"c7e2a035-7568-4455-a3bb-329dcf3276ca\" (UID: \"c7e2a035-7568-4455-a3bb-329dcf3276ca\") " Dec 09 10:11:32 crc kubenswrapper[4824]: I1209 10:11:32.177343 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-scripts\") pod \"c7e2a035-7568-4455-a3bb-329dcf3276ca\" (UID: \"c7e2a035-7568-4455-a3bb-329dcf3276ca\") " Dec 09 10:11:32 crc kubenswrapper[4824]: I1209 10:11:32.186624 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c7e2a035-7568-4455-a3bb-329dcf3276ca" (UID: "c7e2a035-7568-4455-a3bb-329dcf3276ca"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:11:32 crc kubenswrapper[4824]: I1209 10:11:32.187654 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c7e2a035-7568-4455-a3bb-329dcf3276ca" (UID: "c7e2a035-7568-4455-a3bb-329dcf3276ca"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:11:32 crc kubenswrapper[4824]: I1209 10:11:32.188887 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-scripts" (OuterVolumeSpecName: "scripts") pod "c7e2a035-7568-4455-a3bb-329dcf3276ca" (UID: "c7e2a035-7568-4455-a3bb-329dcf3276ca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:11:32 crc kubenswrapper[4824]: I1209 10:11:32.189565 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7e2a035-7568-4455-a3bb-329dcf3276ca-kube-api-access-kwhzk" (OuterVolumeSpecName: "kube-api-access-kwhzk") pod "c7e2a035-7568-4455-a3bb-329dcf3276ca" (UID: "c7e2a035-7568-4455-a3bb-329dcf3276ca"). InnerVolumeSpecName "kube-api-access-kwhzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:11:32 crc kubenswrapper[4824]: I1209 10:11:32.222893 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-config-data" (OuterVolumeSpecName: "config-data") pod "c7e2a035-7568-4455-a3bb-329dcf3276ca" (UID: "c7e2a035-7568-4455-a3bb-329dcf3276ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:11:32 crc kubenswrapper[4824]: I1209 10:11:32.230232 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c7e2a035-7568-4455-a3bb-329dcf3276ca" (UID: "c7e2a035-7568-4455-a3bb-329dcf3276ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:11:32 crc kubenswrapper[4824]: I1209 10:11:32.280626 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwhzk\" (UniqueName: \"kubernetes.io/projected/c7e2a035-7568-4455-a3bb-329dcf3276ca-kube-api-access-kwhzk\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:32 crc kubenswrapper[4824]: I1209 10:11:32.280681 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:32 crc kubenswrapper[4824]: I1209 10:11:32.280696 4824 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:32 crc kubenswrapper[4824]: I1209 10:11:32.280707 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:32 crc kubenswrapper[4824]: I1209 10:11:32.280718 4824 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:32 crc kubenswrapper[4824]: I1209 10:11:32.280731 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7e2a035-7568-4455-a3bb-329dcf3276ca-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:32 crc kubenswrapper[4824]: I1209 10:11:32.472498 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6lhrx" event={"ID":"c7e2a035-7568-4455-a3bb-329dcf3276ca","Type":"ContainerDied","Data":"aa8abf351d6d714313c098612fd6fe9873e75ba590ea98dc826abdd71cb9a285"} Dec 09 10:11:32 crc kubenswrapper[4824]: I1209 10:11:32.472540 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa8abf351d6d714313c098612fd6fe9873e75ba590ea98dc826abdd71cb9a285" Dec 09 10:11:32 crc kubenswrapper[4824]: I1209 10:11:32.472597 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6lhrx" Dec 09 10:11:32 crc kubenswrapper[4824]: I1209 10:11:32.860828 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:11:32 crc kubenswrapper[4824]: I1209 10:11:32.860897 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.231249 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-6lhrx"] Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.244139 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-6lhrx"] Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.333492 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-mlcsg"] Dec 09 10:11:33 crc kubenswrapper[4824]: E1209 10:11:33.334617 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7e2a035-7568-4455-a3bb-329dcf3276ca" containerName="keystone-bootstrap" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.334640 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7e2a035-7568-4455-a3bb-329dcf3276ca" containerName="keystone-bootstrap" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.335069 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7e2a035-7568-4455-a3bb-329dcf3276ca" containerName="keystone-bootstrap" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.336516 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mlcsg" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.339195 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ggbf4" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.340178 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.341217 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.342796 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.343738 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.345605 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mlcsg"] Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.511549 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-fernet-keys\") pod \"keystone-bootstrap-mlcsg\" (UID: \"6a07abd4-056e-4ea7-962f-60f5a03ea701\") " pod="openstack/keystone-bootstrap-mlcsg" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.511918 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-config-data\") pod \"keystone-bootstrap-mlcsg\" (UID: \"6a07abd4-056e-4ea7-962f-60f5a03ea701\") " pod="openstack/keystone-bootstrap-mlcsg" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.512045 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mvsf\" (UniqueName: \"kubernetes.io/projected/6a07abd4-056e-4ea7-962f-60f5a03ea701-kube-api-access-5mvsf\") pod \"keystone-bootstrap-mlcsg\" (UID: \"6a07abd4-056e-4ea7-962f-60f5a03ea701\") " pod="openstack/keystone-bootstrap-mlcsg" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.512136 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-combined-ca-bundle\") pod \"keystone-bootstrap-mlcsg\" (UID: \"6a07abd4-056e-4ea7-962f-60f5a03ea701\") " pod="openstack/keystone-bootstrap-mlcsg" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.512185 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-scripts\") pod \"keystone-bootstrap-mlcsg\" (UID: \"6a07abd4-056e-4ea7-962f-60f5a03ea701\") " pod="openstack/keystone-bootstrap-mlcsg" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.512462 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-credential-keys\") pod \"keystone-bootstrap-mlcsg\" (UID: \"6a07abd4-056e-4ea7-962f-60f5a03ea701\") " pod="openstack/keystone-bootstrap-mlcsg" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.614339 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-fernet-keys\") pod \"keystone-bootstrap-mlcsg\" (UID: \"6a07abd4-056e-4ea7-962f-60f5a03ea701\") " pod="openstack/keystone-bootstrap-mlcsg" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.614512 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-config-data\") pod \"keystone-bootstrap-mlcsg\" (UID: \"6a07abd4-056e-4ea7-962f-60f5a03ea701\") " pod="openstack/keystone-bootstrap-mlcsg" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.614572 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mvsf\" (UniqueName: \"kubernetes.io/projected/6a07abd4-056e-4ea7-962f-60f5a03ea701-kube-api-access-5mvsf\") pod \"keystone-bootstrap-mlcsg\" (UID: \"6a07abd4-056e-4ea7-962f-60f5a03ea701\") " pod="openstack/keystone-bootstrap-mlcsg" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.614622 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-combined-ca-bundle\") pod \"keystone-bootstrap-mlcsg\" (UID: \"6a07abd4-056e-4ea7-962f-60f5a03ea701\") " pod="openstack/keystone-bootstrap-mlcsg" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.614669 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-scripts\") pod \"keystone-bootstrap-mlcsg\" (UID: \"6a07abd4-056e-4ea7-962f-60f5a03ea701\") " pod="openstack/keystone-bootstrap-mlcsg" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.614852 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-credential-keys\") pod \"keystone-bootstrap-mlcsg\" (UID: \"6a07abd4-056e-4ea7-962f-60f5a03ea701\") " pod="openstack/keystone-bootstrap-mlcsg" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.622091 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-fernet-keys\") pod \"keystone-bootstrap-mlcsg\" (UID: \"6a07abd4-056e-4ea7-962f-60f5a03ea701\") " pod="openstack/keystone-bootstrap-mlcsg" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.622203 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-combined-ca-bundle\") pod \"keystone-bootstrap-mlcsg\" (UID: \"6a07abd4-056e-4ea7-962f-60f5a03ea701\") " pod="openstack/keystone-bootstrap-mlcsg" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.623952 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-credential-keys\") pod \"keystone-bootstrap-mlcsg\" (UID: \"6a07abd4-056e-4ea7-962f-60f5a03ea701\") " pod="openstack/keystone-bootstrap-mlcsg" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.648410 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-scripts\") pod \"keystone-bootstrap-mlcsg\" (UID: \"6a07abd4-056e-4ea7-962f-60f5a03ea701\") " pod="openstack/keystone-bootstrap-mlcsg" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.649194 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-config-data\") pod \"keystone-bootstrap-mlcsg\" (UID: \"6a07abd4-056e-4ea7-962f-60f5a03ea701\") " pod="openstack/keystone-bootstrap-mlcsg" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.654292 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mvsf\" (UniqueName: \"kubernetes.io/projected/6a07abd4-056e-4ea7-962f-60f5a03ea701-kube-api-access-5mvsf\") pod \"keystone-bootstrap-mlcsg\" (UID: \"6a07abd4-056e-4ea7-962f-60f5a03ea701\") " pod="openstack/keystone-bootstrap-mlcsg" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.671764 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mlcsg" Dec 09 10:11:33 crc kubenswrapper[4824]: I1209 10:11:33.928013 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7e2a035-7568-4455-a3bb-329dcf3276ca" path="/var/lib/kubelet/pods/c7e2a035-7568-4455-a3bb-329dcf3276ca/volumes" Dec 09 10:11:39 crc kubenswrapper[4824]: E1209 10:11:39.143068 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 09 10:11:39 crc kubenswrapper[4824]: E1209 10:11:39.144183 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mh2p2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-d4klr_openstack(d3d076df-d1e1-473e-9584-79b30cb6d023): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:11:39 crc kubenswrapper[4824]: E1209 10:11:39.145414 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-d4klr" podUID="d3d076df-d1e1-473e-9584-79b30cb6d023" Dec 09 10:11:39 crc kubenswrapper[4824]: I1209 10:11:39.527749 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" podUID="6a45d593-c64c-471e-b0fa-171446548a4d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.168:5353: i/o timeout" Dec 09 10:11:39 crc kubenswrapper[4824]: E1209 10:11:39.583272 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-d4klr" podUID="d3d076df-d1e1-473e-9584-79b30cb6d023" Dec 09 10:11:39 crc kubenswrapper[4824]: E1209 10:11:39.585437 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 09 10:11:39 crc kubenswrapper[4824]: E1209 10:11:39.585618 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndch5f5h677h5d6h598h695h664h5b8h644h686h674h58hc9hd4h665h686h66dh65dh64fh79h54dh74h68bh659h74h565hb8h65fhdch64dh684h7fq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r6zsx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(232c4238-3f59-4e05-ac75-8746ab0a0069): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:11:43 crc kubenswrapper[4824]: I1209 10:11:43.625652 4824 generic.go:334] "Generic (PLEG): container finished" podID="8eb90797-2763-4ffb-8fa5-32a1c8db5c84" containerID="481598142ccbbb37a8c1d4b3c9c41b36e6797b272a9cd42c11a24be6d054d4d6" exitCode=0 Dec 09 10:11:43 crc kubenswrapper[4824]: I1209 10:11:43.625745 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mqjtv" event={"ID":"8eb90797-2763-4ffb-8fa5-32a1c8db5c84","Type":"ContainerDied","Data":"481598142ccbbb37a8c1d4b3c9c41b36e6797b272a9cd42c11a24be6d054d4d6"} Dec 09 10:11:44 crc kubenswrapper[4824]: I1209 10:11:44.529700 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" podUID="6a45d593-c64c-471e-b0fa-171446548a4d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.168:5353: i/o timeout" Dec 09 10:11:44 crc kubenswrapper[4824]: I1209 10:11:44.530497 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:11:47 crc kubenswrapper[4824]: I1209 10:11:47.661542 4824 scope.go:117] "RemoveContainer" containerID="270a015831ebac9b4ca1a30733ca8ef7d77be57e1e8a715249e2282449121282" Dec 09 10:11:47 crc kubenswrapper[4824]: I1209 10:11:47.938216 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.053886 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q44x9\" (UniqueName: \"kubernetes.io/projected/6a45d593-c64c-471e-b0fa-171446548a4d-kube-api-access-q44x9\") pod \"6a45d593-c64c-471e-b0fa-171446548a4d\" (UID: \"6a45d593-c64c-471e-b0fa-171446548a4d\") " Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.053994 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-config\") pod \"6a45d593-c64c-471e-b0fa-171446548a4d\" (UID: \"6a45d593-c64c-471e-b0fa-171446548a4d\") " Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.054037 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-dns-svc\") pod \"6a45d593-c64c-471e-b0fa-171446548a4d\" (UID: \"6a45d593-c64c-471e-b0fa-171446548a4d\") " Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.054063 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-ovsdbserver-nb\") pod \"6a45d593-c64c-471e-b0fa-171446548a4d\" (UID: \"6a45d593-c64c-471e-b0fa-171446548a4d\") " Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.054090 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-ovsdbserver-sb\") pod \"6a45d593-c64c-471e-b0fa-171446548a4d\" (UID: \"6a45d593-c64c-471e-b0fa-171446548a4d\") " Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.054110 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-dns-swift-storage-0\") pod \"6a45d593-c64c-471e-b0fa-171446548a4d\" (UID: \"6a45d593-c64c-471e-b0fa-171446548a4d\") " Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.110658 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a45d593-c64c-471e-b0fa-171446548a4d-kube-api-access-q44x9" (OuterVolumeSpecName: "kube-api-access-q44x9") pod "6a45d593-c64c-471e-b0fa-171446548a4d" (UID: "6a45d593-c64c-471e-b0fa-171446548a4d"). InnerVolumeSpecName "kube-api-access-q44x9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.158331 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q44x9\" (UniqueName: \"kubernetes.io/projected/6a45d593-c64c-471e-b0fa-171446548a4d-kube-api-access-q44x9\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.233568 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6a45d593-c64c-471e-b0fa-171446548a4d" (UID: "6a45d593-c64c-471e-b0fa-171446548a4d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.239239 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6a45d593-c64c-471e-b0fa-171446548a4d" (UID: "6a45d593-c64c-471e-b0fa-171446548a4d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.249919 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6a45d593-c64c-471e-b0fa-171446548a4d" (UID: "6a45d593-c64c-471e-b0fa-171446548a4d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.260381 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.260421 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.260435 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.282932 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-config" (OuterVolumeSpecName: "config") pod "6a45d593-c64c-471e-b0fa-171446548a4d" (UID: "6a45d593-c64c-471e-b0fa-171446548a4d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.330520 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6a45d593-c64c-471e-b0fa-171446548a4d" (UID: "6a45d593-c64c-471e-b0fa-171446548a4d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.362838 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.362870 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a45d593-c64c-471e-b0fa-171446548a4d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.513745 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mqjtv" Dec 09 10:11:48 crc kubenswrapper[4824]: E1209 10:11:48.529534 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified" Dec 09 10:11:48 crc kubenswrapper[4824]: E1209 10:11:48.529720 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p2csn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-vfnsw_openstack(405d0a5e-fd62-4146-bfb9-96ff6c077836): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:11:48 crc kubenswrapper[4824]: E1209 10:11:48.531470 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-vfnsw" podUID="405d0a5e-fd62-4146-bfb9-96ff6c077836" Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.566020 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8eb90797-2763-4ffb-8fa5-32a1c8db5c84-combined-ca-bundle\") pod \"8eb90797-2763-4ffb-8fa5-32a1c8db5c84\" (UID: \"8eb90797-2763-4ffb-8fa5-32a1c8db5c84\") " Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.598439 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8eb90797-2763-4ffb-8fa5-32a1c8db5c84-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8eb90797-2763-4ffb-8fa5-32a1c8db5c84" (UID: "8eb90797-2763-4ffb-8fa5-32a1c8db5c84"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.667450 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8eb90797-2763-4ffb-8fa5-32a1c8db5c84-config\") pod \"8eb90797-2763-4ffb-8fa5-32a1c8db5c84\" (UID: \"8eb90797-2763-4ffb-8fa5-32a1c8db5c84\") " Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.667763 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rr4gr\" (UniqueName: \"kubernetes.io/projected/8eb90797-2763-4ffb-8fa5-32a1c8db5c84-kube-api-access-rr4gr\") pod \"8eb90797-2763-4ffb-8fa5-32a1c8db5c84\" (UID: \"8eb90797-2763-4ffb-8fa5-32a1c8db5c84\") " Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.668802 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8eb90797-2763-4ffb-8fa5-32a1c8db5c84-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.672711 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8eb90797-2763-4ffb-8fa5-32a1c8db5c84-kube-api-access-rr4gr" (OuterVolumeSpecName: "kube-api-access-rr4gr") pod "8eb90797-2763-4ffb-8fa5-32a1c8db5c84" (UID: "8eb90797-2763-4ffb-8fa5-32a1c8db5c84"). InnerVolumeSpecName "kube-api-access-rr4gr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.691664 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" event={"ID":"6a45d593-c64c-471e-b0fa-171446548a4d","Type":"ContainerDied","Data":"182446cf60766ff1ce3f29478da994c779b28b093149949b54447f66562c814f"} Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.691800 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.698930 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mqjtv" event={"ID":"8eb90797-2763-4ffb-8fa5-32a1c8db5c84","Type":"ContainerDied","Data":"ae2b0b9451945375a741db97237b099463efb38f3e2a7f706dc2ee68147ed469"} Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.698976 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae2b0b9451945375a741db97237b099463efb38f3e2a7f706dc2ee68147ed469" Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.698943 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mqjtv" Dec 09 10:11:48 crc kubenswrapper[4824]: E1209 10:11:48.702568 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified\\\"\"" pod="openstack/heat-db-sync-vfnsw" podUID="405d0a5e-fd62-4146-bfb9-96ff6c077836" Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.722173 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8eb90797-2763-4ffb-8fa5-32a1c8db5c84-config" (OuterVolumeSpecName: "config") pod "8eb90797-2763-4ffb-8fa5-32a1c8db5c84" (UID: "8eb90797-2763-4ffb-8fa5-32a1c8db5c84"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.770983 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rr4gr\" (UniqueName: \"kubernetes.io/projected/8eb90797-2763-4ffb-8fa5-32a1c8db5c84-kube-api-access-rr4gr\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.771029 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8eb90797-2763-4ffb-8fa5-32a1c8db5c84-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.829836 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-fgj5w"] Dec 09 10:11:48 crc kubenswrapper[4824]: I1209 10:11:48.840492 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-fgj5w"] Dec 09 10:11:49 crc kubenswrapper[4824]: I1209 10:11:49.530717 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-fgj5w" podUID="6a45d593-c64c-471e-b0fa-171446548a4d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.168:5353: i/o timeout" Dec 09 10:11:49 crc kubenswrapper[4824]: I1209 10:11:49.708646 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-pmxwc"] Dec 09 10:11:49 crc kubenswrapper[4824]: E1209 10:11:49.771905 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a45d593-c64c-471e-b0fa-171446548a4d" containerName="init" Dec 09 10:11:49 crc kubenswrapper[4824]: I1209 10:11:49.771952 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a45d593-c64c-471e-b0fa-171446548a4d" containerName="init" Dec 09 10:11:49 crc kubenswrapper[4824]: E1209 10:11:49.771998 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8eb90797-2763-4ffb-8fa5-32a1c8db5c84" containerName="neutron-db-sync" Dec 09 10:11:49 crc kubenswrapper[4824]: I1209 10:11:49.772008 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8eb90797-2763-4ffb-8fa5-32a1c8db5c84" containerName="neutron-db-sync" Dec 09 10:11:49 crc kubenswrapper[4824]: E1209 10:11:49.772029 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a45d593-c64c-471e-b0fa-171446548a4d" containerName="dnsmasq-dns" Dec 09 10:11:49 crc kubenswrapper[4824]: I1209 10:11:49.772040 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a45d593-c64c-471e-b0fa-171446548a4d" containerName="dnsmasq-dns" Dec 09 10:11:49 crc kubenswrapper[4824]: I1209 10:11:49.772488 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8eb90797-2763-4ffb-8fa5-32a1c8db5c84" containerName="neutron-db-sync" Dec 09 10:11:49 crc kubenswrapper[4824]: I1209 10:11:49.772512 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a45d593-c64c-471e-b0fa-171446548a4d" containerName="dnsmasq-dns" Dec 09 10:11:49 crc kubenswrapper[4824]: I1209 10:11:49.774487 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" Dec 09 10:11:49 crc kubenswrapper[4824]: I1209 10:11:49.875638 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-pmxwc"] Dec 09 10:11:49 crc kubenswrapper[4824]: I1209 10:11:49.958856 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-pmxwc\" (UID: \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\") " pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" Dec 09 10:11:49 crc kubenswrapper[4824]: I1209 10:11:49.959384 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-pmxwc\" (UID: \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\") " pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" Dec 09 10:11:49 crc kubenswrapper[4824]: I1209 10:11:49.959673 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-config\") pod \"dnsmasq-dns-55f844cf75-pmxwc\" (UID: \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\") " pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" Dec 09 10:11:49 crc kubenswrapper[4824]: I1209 10:11:49.959841 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-pmxwc\" (UID: \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\") " pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" Dec 09 10:11:49 crc kubenswrapper[4824]: I1209 10:11:49.960208 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-dns-svc\") pod \"dnsmasq-dns-55f844cf75-pmxwc\" (UID: \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\") " pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" Dec 09 10:11:49 crc kubenswrapper[4824]: I1209 10:11:49.964339 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a45d593-c64c-471e-b0fa-171446548a4d" path="/var/lib/kubelet/pods/6a45d593-c64c-471e-b0fa-171446548a4d/volumes" Dec 09 10:11:49 crc kubenswrapper[4824]: I1209 10:11:49.971160 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzwkk\" (UniqueName: \"kubernetes.io/projected/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-kube-api-access-tzwkk\") pod \"dnsmasq-dns-55f844cf75-pmxwc\" (UID: \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\") " pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" Dec 09 10:11:49 crc kubenswrapper[4824]: I1209 10:11:49.979468 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5855dddcbd-6qlwr"] Dec 09 10:11:49 crc kubenswrapper[4824]: I1209 10:11:49.982125 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5855dddcbd-6qlwr" Dec 09 10:11:49 crc kubenswrapper[4824]: I1209 10:11:49.990933 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 09 10:11:49 crc kubenswrapper[4824]: I1209 10:11:49.991009 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-r4m6h" Dec 09 10:11:49 crc kubenswrapper[4824]: I1209 10:11:49.991153 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 09 10:11:49 crc kubenswrapper[4824]: I1209 10:11:49.993500 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.007137 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5855dddcbd-6qlwr"] Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.073790 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-dns-svc\") pod \"dnsmasq-dns-55f844cf75-pmxwc\" (UID: \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\") " pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.073853 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzwkk\" (UniqueName: \"kubernetes.io/projected/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-kube-api-access-tzwkk\") pod \"dnsmasq-dns-55f844cf75-pmxwc\" (UID: \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\") " pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.073891 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-combined-ca-bundle\") pod \"neutron-5855dddcbd-6qlwr\" (UID: \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\") " pod="openstack/neutron-5855dddcbd-6qlwr" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.073923 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-pmxwc\" (UID: \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\") " pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.074042 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-pmxwc\" (UID: \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\") " pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.074115 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-httpd-config\") pod \"neutron-5855dddcbd-6qlwr\" (UID: \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\") " pod="openstack/neutron-5855dddcbd-6qlwr" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.074159 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-config\") pod \"neutron-5855dddcbd-6qlwr\" (UID: \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\") " pod="openstack/neutron-5855dddcbd-6qlwr" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.074188 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgxh9\" (UniqueName: \"kubernetes.io/projected/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-kube-api-access-pgxh9\") pod \"neutron-5855dddcbd-6qlwr\" (UID: \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\") " pod="openstack/neutron-5855dddcbd-6qlwr" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.074237 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-config\") pod \"dnsmasq-dns-55f844cf75-pmxwc\" (UID: \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\") " pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.074277 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-pmxwc\" (UID: \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\") " pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.074293 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-ovndb-tls-certs\") pod \"neutron-5855dddcbd-6qlwr\" (UID: \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\") " pod="openstack/neutron-5855dddcbd-6qlwr" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.075273 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-pmxwc\" (UID: \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\") " pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.075772 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-pmxwc\" (UID: \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\") " pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.075845 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-dns-svc\") pod \"dnsmasq-dns-55f844cf75-pmxwc\" (UID: \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\") " pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.075872 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-pmxwc\" (UID: \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\") " pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.075966 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-config\") pod \"dnsmasq-dns-55f844cf75-pmxwc\" (UID: \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\") " pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.099551 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzwkk\" (UniqueName: \"kubernetes.io/projected/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-kube-api-access-tzwkk\") pod \"dnsmasq-dns-55f844cf75-pmxwc\" (UID: \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\") " pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.176514 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-combined-ca-bundle\") pod \"neutron-5855dddcbd-6qlwr\" (UID: \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\") " pod="openstack/neutron-5855dddcbd-6qlwr" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.177147 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-httpd-config\") pod \"neutron-5855dddcbd-6qlwr\" (UID: \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\") " pod="openstack/neutron-5855dddcbd-6qlwr" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.177179 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-config\") pod \"neutron-5855dddcbd-6qlwr\" (UID: \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\") " pod="openstack/neutron-5855dddcbd-6qlwr" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.177198 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgxh9\" (UniqueName: \"kubernetes.io/projected/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-kube-api-access-pgxh9\") pod \"neutron-5855dddcbd-6qlwr\" (UID: \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\") " pod="openstack/neutron-5855dddcbd-6qlwr" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.177247 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-ovndb-tls-certs\") pod \"neutron-5855dddcbd-6qlwr\" (UID: \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\") " pod="openstack/neutron-5855dddcbd-6qlwr" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.181528 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.182178 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-combined-ca-bundle\") pod \"neutron-5855dddcbd-6qlwr\" (UID: \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\") " pod="openstack/neutron-5855dddcbd-6qlwr" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.184683 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-httpd-config\") pod \"neutron-5855dddcbd-6qlwr\" (UID: \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\") " pod="openstack/neutron-5855dddcbd-6qlwr" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.186912 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-ovndb-tls-certs\") pod \"neutron-5855dddcbd-6qlwr\" (UID: \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\") " pod="openstack/neutron-5855dddcbd-6qlwr" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.187991 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-config\") pod \"neutron-5855dddcbd-6qlwr\" (UID: \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\") " pod="openstack/neutron-5855dddcbd-6qlwr" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.201956 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgxh9\" (UniqueName: \"kubernetes.io/projected/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-kube-api-access-pgxh9\") pod \"neutron-5855dddcbd-6qlwr\" (UID: \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\") " pod="openstack/neutron-5855dddcbd-6qlwr" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.324156 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5855dddcbd-6qlwr" Dec 09 10:11:50 crc kubenswrapper[4824]: E1209 10:11:50.326723 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 09 10:11:50 crc kubenswrapper[4824]: E1209 10:11:50.326972 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lwnjz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-jghjq_openstack(dbceafb7-272a-45fc-9b56-df631986f90a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:11:50 crc kubenswrapper[4824]: E1209 10:11:50.328617 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-jghjq" podUID="dbceafb7-272a-45fc-9b56-df631986f90a" Dec 09 10:11:50 crc kubenswrapper[4824]: I1209 10:11:50.421089 4824 scope.go:117] "RemoveContainer" containerID="085feec377c96e6eba63a801f66855826edb5591113a6501dd0576d7f8ea05d0" Dec 09 10:11:50 crc kubenswrapper[4824]: E1209 10:11:50.803400 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-jghjq" podUID="dbceafb7-272a-45fc-9b56-df631986f90a" Dec 09 10:11:51 crc kubenswrapper[4824]: I1209 10:11:51.205255 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.604415 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7d56cf488f-6wtnd"] Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.607122 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.609316 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.616290 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7d56cf488f-6wtnd"] Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.621448 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.631637 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef937030-1ed4-4462-b362-aa335e77a9f6-internal-tls-certs\") pod \"neutron-7d56cf488f-6wtnd\" (UID: \"ef937030-1ed4-4462-b362-aa335e77a9f6\") " pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.631954 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ef937030-1ed4-4462-b362-aa335e77a9f6-config\") pod \"neutron-7d56cf488f-6wtnd\" (UID: \"ef937030-1ed4-4462-b362-aa335e77a9f6\") " pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.632194 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef937030-1ed4-4462-b362-aa335e77a9f6-public-tls-certs\") pod \"neutron-7d56cf488f-6wtnd\" (UID: \"ef937030-1ed4-4462-b362-aa335e77a9f6\") " pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.632407 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef937030-1ed4-4462-b362-aa335e77a9f6-ovndb-tls-certs\") pod \"neutron-7d56cf488f-6wtnd\" (UID: \"ef937030-1ed4-4462-b362-aa335e77a9f6\") " pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.632486 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f7k9\" (UniqueName: \"kubernetes.io/projected/ef937030-1ed4-4462-b362-aa335e77a9f6-kube-api-access-6f7k9\") pod \"neutron-7d56cf488f-6wtnd\" (UID: \"ef937030-1ed4-4462-b362-aa335e77a9f6\") " pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.632536 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef937030-1ed4-4462-b362-aa335e77a9f6-combined-ca-bundle\") pod \"neutron-7d56cf488f-6wtnd\" (UID: \"ef937030-1ed4-4462-b362-aa335e77a9f6\") " pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.632585 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ef937030-1ed4-4462-b362-aa335e77a9f6-httpd-config\") pod \"neutron-7d56cf488f-6wtnd\" (UID: \"ef937030-1ed4-4462-b362-aa335e77a9f6\") " pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.735398 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ef937030-1ed4-4462-b362-aa335e77a9f6-config\") pod \"neutron-7d56cf488f-6wtnd\" (UID: \"ef937030-1ed4-4462-b362-aa335e77a9f6\") " pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.735580 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef937030-1ed4-4462-b362-aa335e77a9f6-public-tls-certs\") pod \"neutron-7d56cf488f-6wtnd\" (UID: \"ef937030-1ed4-4462-b362-aa335e77a9f6\") " pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.735757 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef937030-1ed4-4462-b362-aa335e77a9f6-ovndb-tls-certs\") pod \"neutron-7d56cf488f-6wtnd\" (UID: \"ef937030-1ed4-4462-b362-aa335e77a9f6\") " pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.735860 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f7k9\" (UniqueName: \"kubernetes.io/projected/ef937030-1ed4-4462-b362-aa335e77a9f6-kube-api-access-6f7k9\") pod \"neutron-7d56cf488f-6wtnd\" (UID: \"ef937030-1ed4-4462-b362-aa335e77a9f6\") " pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.735915 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef937030-1ed4-4462-b362-aa335e77a9f6-combined-ca-bundle\") pod \"neutron-7d56cf488f-6wtnd\" (UID: \"ef937030-1ed4-4462-b362-aa335e77a9f6\") " pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.735951 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ef937030-1ed4-4462-b362-aa335e77a9f6-httpd-config\") pod \"neutron-7d56cf488f-6wtnd\" (UID: \"ef937030-1ed4-4462-b362-aa335e77a9f6\") " pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.736035 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef937030-1ed4-4462-b362-aa335e77a9f6-internal-tls-certs\") pod \"neutron-7d56cf488f-6wtnd\" (UID: \"ef937030-1ed4-4462-b362-aa335e77a9f6\") " pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.744997 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ef937030-1ed4-4462-b362-aa335e77a9f6-httpd-config\") pod \"neutron-7d56cf488f-6wtnd\" (UID: \"ef937030-1ed4-4462-b362-aa335e77a9f6\") " pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.745017 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ef937030-1ed4-4462-b362-aa335e77a9f6-config\") pod \"neutron-7d56cf488f-6wtnd\" (UID: \"ef937030-1ed4-4462-b362-aa335e77a9f6\") " pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.744998 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef937030-1ed4-4462-b362-aa335e77a9f6-internal-tls-certs\") pod \"neutron-7d56cf488f-6wtnd\" (UID: \"ef937030-1ed4-4462-b362-aa335e77a9f6\") " pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.745546 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef937030-1ed4-4462-b362-aa335e77a9f6-ovndb-tls-certs\") pod \"neutron-7d56cf488f-6wtnd\" (UID: \"ef937030-1ed4-4462-b362-aa335e77a9f6\") " pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.745860 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef937030-1ed4-4462-b362-aa335e77a9f6-combined-ca-bundle\") pod \"neutron-7d56cf488f-6wtnd\" (UID: \"ef937030-1ed4-4462-b362-aa335e77a9f6\") " pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.746366 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef937030-1ed4-4462-b362-aa335e77a9f6-public-tls-certs\") pod \"neutron-7d56cf488f-6wtnd\" (UID: \"ef937030-1ed4-4462-b362-aa335e77a9f6\") " pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.756039 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f7k9\" (UniqueName: \"kubernetes.io/projected/ef937030-1ed4-4462-b362-aa335e77a9f6-kube-api-access-6f7k9\") pod \"neutron-7d56cf488f-6wtnd\" (UID: \"ef937030-1ed4-4462-b362-aa335e77a9f6\") " pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:11:52 crc kubenswrapper[4824]: I1209 10:11:52.932436 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:11:54 crc kubenswrapper[4824]: W1209 10:11:54.773041 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod988197de_9531_4cae_92d0_ec7f836a8c73.slice/crio-6f3e45ac8a89ab64150f1214fbb1c24d6bbd24b99cf2811f5f28ec7c5d7d341b WatchSource:0}: Error finding container 6f3e45ac8a89ab64150f1214fbb1c24d6bbd24b99cf2811f5f28ec7c5d7d341b: Status 404 returned error can't find the container with id 6f3e45ac8a89ab64150f1214fbb1c24d6bbd24b99cf2811f5f28ec7c5d7d341b Dec 09 10:11:54 crc kubenswrapper[4824]: I1209 10:11:54.789159 4824 scope.go:117] "RemoveContainer" containerID="0de4dfe5dc112545d3c8f9f2f646791151f090c0f7071a68110442d9a180004b" Dec 09 10:11:54 crc kubenswrapper[4824]: I1209 10:11:54.874174 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"988197de-9531-4cae-92d0-ec7f836a8c73","Type":"ContainerStarted","Data":"6f3e45ac8a89ab64150f1214fbb1c24d6bbd24b99cf2811f5f28ec7c5d7d341b"} Dec 09 10:11:55 crc kubenswrapper[4824]: I1209 10:11:55.489083 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 10:11:56 crc kubenswrapper[4824]: I1209 10:11:56.014894 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-pmxwc"] Dec 09 10:11:56 crc kubenswrapper[4824]: I1209 10:11:56.031361 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mlcsg"] Dec 09 10:11:56 crc kubenswrapper[4824]: I1209 10:11:56.069981 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d974a3d8-22a8-4301-bb5f-28011189b4dc","Type":"ContainerStarted","Data":"f2bdb66743ef7568d69f8bfe9763eeb74d67d2d41cadafb0196db3222cc2275d"} Dec 09 10:11:56 crc kubenswrapper[4824]: I1209 10:11:56.126763 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ql4lt" event={"ID":"aae4e0f1-0295-4964-8b34-56bdab58a152","Type":"ContainerStarted","Data":"3a153c578db1e11c3e8a67688b76a866e98ce0c6d3d6ff91aca0a32828d69f90"} Dec 09 10:11:56 crc kubenswrapper[4824]: I1209 10:11:56.137228 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-d4klr" event={"ID":"d3d076df-d1e1-473e-9584-79b30cb6d023","Type":"ContainerStarted","Data":"66df44334c000ef2419a325e951be6d72b85973d4249b0d527f40d2ad0d6338b"} Dec 09 10:11:56 crc kubenswrapper[4824]: I1209 10:11:56.146362 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5855dddcbd-6qlwr"] Dec 09 10:11:56 crc kubenswrapper[4824]: I1209 10:11:56.155692 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"232c4238-3f59-4e05-ac75-8746ab0a0069","Type":"ContainerStarted","Data":"89110daf7562ee61890832514845441d50bf7e08943934df95b78cd71224c749"} Dec 09 10:11:56 crc kubenswrapper[4824]: I1209 10:11:56.158853 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-ql4lt" podStartSLOduration=8.979006812 podStartE2EDuration="41.158827797s" podCreationTimestamp="2025-12-09 10:11:15 +0000 UTC" firstStartedPulling="2025-12-09 10:11:18.083828573 +0000 UTC m=+1434.418333240" lastFinishedPulling="2025-12-09 10:11:50.263649558 +0000 UTC m=+1466.598154225" observedRunningTime="2025-12-09 10:11:56.153612291 +0000 UTC m=+1472.488116968" watchObservedRunningTime="2025-12-09 10:11:56.158827797 +0000 UTC m=+1472.493332464" Dec 09 10:11:56 crc kubenswrapper[4824]: I1209 10:11:56.213246 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-d4klr" podStartSLOduration=4.961393705 podStartE2EDuration="42.213219909s" podCreationTimestamp="2025-12-09 10:11:14 +0000 UTC" firstStartedPulling="2025-12-09 10:11:17.927832497 +0000 UTC m=+1434.262337164" lastFinishedPulling="2025-12-09 10:11:55.179658701 +0000 UTC m=+1471.514163368" observedRunningTime="2025-12-09 10:11:56.170249211 +0000 UTC m=+1472.504753888" watchObservedRunningTime="2025-12-09 10:11:56.213219909 +0000 UTC m=+1472.547724576" Dec 09 10:11:57 crc kubenswrapper[4824]: I1209 10:11:57.053305 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7d56cf488f-6wtnd"] Dec 09 10:11:57 crc kubenswrapper[4824]: W1209 10:11:57.081536 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef937030_1ed4_4462_b362_aa335e77a9f6.slice/crio-90082a7265bdbb1025a3909f61c653d2ffc51cbf24f16e1942e788a9988b19a4 WatchSource:0}: Error finding container 90082a7265bdbb1025a3909f61c653d2ffc51cbf24f16e1942e788a9988b19a4: Status 404 returned error can't find the container with id 90082a7265bdbb1025a3909f61c653d2ffc51cbf24f16e1942e788a9988b19a4 Dec 09 10:11:57 crc kubenswrapper[4824]: I1209 10:11:57.211382 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d974a3d8-22a8-4301-bb5f-28011189b4dc","Type":"ContainerStarted","Data":"ada896b2969c2cfab19365a8fe216e3c2acd2174978d17ef9a24d6c1b93b028d"} Dec 09 10:11:57 crc kubenswrapper[4824]: I1209 10:11:57.215620 4824 generic.go:334] "Generic (PLEG): container finished" podID="ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780" containerID="fa297085f99c182ce1a59f3a61473b47e2dccf05183db5580ccc4e987dfa27be" exitCode=0 Dec 09 10:11:57 crc kubenswrapper[4824]: I1209 10:11:57.215697 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" event={"ID":"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780","Type":"ContainerDied","Data":"fa297085f99c182ce1a59f3a61473b47e2dccf05183db5580ccc4e987dfa27be"} Dec 09 10:11:57 crc kubenswrapper[4824]: I1209 10:11:57.215731 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" event={"ID":"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780","Type":"ContainerStarted","Data":"ca4d28649370aeb222a8941f8d1af79178a3523d1d6245363a4771dde7373742"} Dec 09 10:11:57 crc kubenswrapper[4824]: I1209 10:11:57.230074 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5855dddcbd-6qlwr" event={"ID":"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3","Type":"ContainerStarted","Data":"8e6ba1b57bd352f45b14b90495296a5ad4334e5ba289f62bcfb6b62fdd0b9de2"} Dec 09 10:11:57 crc kubenswrapper[4824]: I1209 10:11:57.230138 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5855dddcbd-6qlwr" event={"ID":"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3","Type":"ContainerStarted","Data":"39ac59a1f6fbf896b4c2d08f9997e40dec30a053bbbcc01bf095c383bce4b7af"} Dec 09 10:11:57 crc kubenswrapper[4824]: I1209 10:11:57.230151 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5855dddcbd-6qlwr" event={"ID":"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3","Type":"ContainerStarted","Data":"06ca3a68ff07de58d1ef52781bf7760d4413ace0eb94f551b127a233be609488"} Dec 09 10:11:57 crc kubenswrapper[4824]: I1209 10:11:57.232016 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5855dddcbd-6qlwr" Dec 09 10:11:57 crc kubenswrapper[4824]: I1209 10:11:57.250822 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"988197de-9531-4cae-92d0-ec7f836a8c73","Type":"ContainerStarted","Data":"a2aedab1b4a2c8aabfcc925c11aee95fa4283be5ad4e6682922565866d35a7c2"} Dec 09 10:11:57 crc kubenswrapper[4824]: I1209 10:11:57.250881 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"988197de-9531-4cae-92d0-ec7f836a8c73","Type":"ContainerStarted","Data":"75fa1185360d66e483556e52bfc7505304668e9f6690cfd56a3faffca30c8311"} Dec 09 10:11:57 crc kubenswrapper[4824]: I1209 10:11:57.265657 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mlcsg" event={"ID":"6a07abd4-056e-4ea7-962f-60f5a03ea701","Type":"ContainerStarted","Data":"e215988c9459664b8ba3af9c71ac7573e5936c4f048333de2977d582be175243"} Dec 09 10:11:57 crc kubenswrapper[4824]: I1209 10:11:57.265708 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mlcsg" event={"ID":"6a07abd4-056e-4ea7-962f-60f5a03ea701","Type":"ContainerStarted","Data":"a4b5f0cd195efa7bf674fc7266a9484f3d9672b8cc14cc0b462b43fd6440895a"} Dec 09 10:11:57 crc kubenswrapper[4824]: I1209 10:11:57.277264 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5855dddcbd-6qlwr" podStartSLOduration=8.277247936 podStartE2EDuration="8.277247936s" podCreationTimestamp="2025-12-09 10:11:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:11:57.264552592 +0000 UTC m=+1473.599057269" watchObservedRunningTime="2025-12-09 10:11:57.277247936 +0000 UTC m=+1473.611752593" Dec 09 10:11:57 crc kubenswrapper[4824]: I1209 10:11:57.277307 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d56cf488f-6wtnd" event={"ID":"ef937030-1ed4-4462-b362-aa335e77a9f6","Type":"ContainerStarted","Data":"90082a7265bdbb1025a3909f61c653d2ffc51cbf24f16e1942e788a9988b19a4"} Dec 09 10:11:57 crc kubenswrapper[4824]: I1209 10:11:57.301246 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=31.301224899 podStartE2EDuration="31.301224899s" podCreationTimestamp="2025-12-09 10:11:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:11:57.289322941 +0000 UTC m=+1473.623827628" watchObservedRunningTime="2025-12-09 10:11:57.301224899 +0000 UTC m=+1473.635729566" Dec 09 10:11:57 crc kubenswrapper[4824]: I1209 10:11:57.333750 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-mlcsg" podStartSLOduration=24.333728374 podStartE2EDuration="24.333728374s" podCreationTimestamp="2025-12-09 10:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:11:57.310096672 +0000 UTC m=+1473.644601339" watchObservedRunningTime="2025-12-09 10:11:57.333728374 +0000 UTC m=+1473.668233051" Dec 09 10:11:57 crc kubenswrapper[4824]: I1209 10:11:57.875603 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lhlfr"] Dec 09 10:11:57 crc kubenswrapper[4824]: I1209 10:11:57.886615 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lhlfr" Dec 09 10:11:57 crc kubenswrapper[4824]: I1209 10:11:57.937587 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lhlfr"] Dec 09 10:11:58 crc kubenswrapper[4824]: I1209 10:11:58.019203 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad8cfdb4-f7e5-4b30-a1b8-a680b111f788-catalog-content\") pod \"redhat-operators-lhlfr\" (UID: \"ad8cfdb4-f7e5-4b30-a1b8-a680b111f788\") " pod="openshift-marketplace/redhat-operators-lhlfr" Dec 09 10:11:58 crc kubenswrapper[4824]: I1209 10:11:58.019342 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nctnx\" (UniqueName: \"kubernetes.io/projected/ad8cfdb4-f7e5-4b30-a1b8-a680b111f788-kube-api-access-nctnx\") pod \"redhat-operators-lhlfr\" (UID: \"ad8cfdb4-f7e5-4b30-a1b8-a680b111f788\") " pod="openshift-marketplace/redhat-operators-lhlfr" Dec 09 10:11:58 crc kubenswrapper[4824]: I1209 10:11:58.019385 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad8cfdb4-f7e5-4b30-a1b8-a680b111f788-utilities\") pod \"redhat-operators-lhlfr\" (UID: \"ad8cfdb4-f7e5-4b30-a1b8-a680b111f788\") " pod="openshift-marketplace/redhat-operators-lhlfr" Dec 09 10:11:58 crc kubenswrapper[4824]: I1209 10:11:58.121497 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad8cfdb4-f7e5-4b30-a1b8-a680b111f788-catalog-content\") pod \"redhat-operators-lhlfr\" (UID: \"ad8cfdb4-f7e5-4b30-a1b8-a680b111f788\") " pod="openshift-marketplace/redhat-operators-lhlfr" Dec 09 10:11:58 crc kubenswrapper[4824]: I1209 10:11:58.121642 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nctnx\" (UniqueName: \"kubernetes.io/projected/ad8cfdb4-f7e5-4b30-a1b8-a680b111f788-kube-api-access-nctnx\") pod \"redhat-operators-lhlfr\" (UID: \"ad8cfdb4-f7e5-4b30-a1b8-a680b111f788\") " pod="openshift-marketplace/redhat-operators-lhlfr" Dec 09 10:11:58 crc kubenswrapper[4824]: I1209 10:11:58.121705 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad8cfdb4-f7e5-4b30-a1b8-a680b111f788-utilities\") pod \"redhat-operators-lhlfr\" (UID: \"ad8cfdb4-f7e5-4b30-a1b8-a680b111f788\") " pod="openshift-marketplace/redhat-operators-lhlfr" Dec 09 10:11:58 crc kubenswrapper[4824]: I1209 10:11:58.122418 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad8cfdb4-f7e5-4b30-a1b8-a680b111f788-catalog-content\") pod \"redhat-operators-lhlfr\" (UID: \"ad8cfdb4-f7e5-4b30-a1b8-a680b111f788\") " pod="openshift-marketplace/redhat-operators-lhlfr" Dec 09 10:11:58 crc kubenswrapper[4824]: I1209 10:11:58.122462 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad8cfdb4-f7e5-4b30-a1b8-a680b111f788-utilities\") pod \"redhat-operators-lhlfr\" (UID: \"ad8cfdb4-f7e5-4b30-a1b8-a680b111f788\") " pod="openshift-marketplace/redhat-operators-lhlfr" Dec 09 10:11:58 crc kubenswrapper[4824]: I1209 10:11:58.143808 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nctnx\" (UniqueName: \"kubernetes.io/projected/ad8cfdb4-f7e5-4b30-a1b8-a680b111f788-kube-api-access-nctnx\") pod \"redhat-operators-lhlfr\" (UID: \"ad8cfdb4-f7e5-4b30-a1b8-a680b111f788\") " pod="openshift-marketplace/redhat-operators-lhlfr" Dec 09 10:11:58 crc kubenswrapper[4824]: I1209 10:11:58.233274 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lhlfr" Dec 09 10:11:58 crc kubenswrapper[4824]: I1209 10:11:58.320703 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d56cf488f-6wtnd" event={"ID":"ef937030-1ed4-4462-b362-aa335e77a9f6","Type":"ContainerStarted","Data":"682a0a822f574d2a0f451abc81e32349f4183fb9e17879cc71c04491682e2642"} Dec 09 10:11:58 crc kubenswrapper[4824]: I1209 10:11:58.321089 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d56cf488f-6wtnd" event={"ID":"ef937030-1ed4-4462-b362-aa335e77a9f6","Type":"ContainerStarted","Data":"e55b493b376c29e9827559c62f5516d938993fd5f061b9bf25834925d74bf2d0"} Dec 09 10:11:58 crc kubenswrapper[4824]: I1209 10:11:58.321151 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:11:58 crc kubenswrapper[4824]: I1209 10:11:58.331712 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d974a3d8-22a8-4301-bb5f-28011189b4dc","Type":"ContainerStarted","Data":"0763c351bab138a499fba2d8c867605e7f3926bfe2ca508f44cd60ce8bb1b6d9"} Dec 09 10:11:58 crc kubenswrapper[4824]: I1209 10:11:58.337210 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" event={"ID":"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780","Type":"ContainerStarted","Data":"20a51d23f331fb43fd5bb76eccbc3e81fdb36fcc151c9f8fb0f57733abeca4bf"} Dec 09 10:11:58 crc kubenswrapper[4824]: I1209 10:11:58.337775 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" Dec 09 10:11:58 crc kubenswrapper[4824]: I1209 10:11:58.398015 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7d56cf488f-6wtnd" podStartSLOduration=6.397982919 podStartE2EDuration="6.397982919s" podCreationTimestamp="2025-12-09 10:11:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:11:58.359481353 +0000 UTC m=+1474.693986040" watchObservedRunningTime="2025-12-09 10:11:58.397982919 +0000 UTC m=+1474.732487576" Dec 09 10:11:58 crc kubenswrapper[4824]: I1209 10:11:58.415031 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=33.41500635 podStartE2EDuration="33.41500635s" podCreationTimestamp="2025-12-09 10:11:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:11:58.40648395 +0000 UTC m=+1474.740988647" watchObservedRunningTime="2025-12-09 10:11:58.41500635 +0000 UTC m=+1474.749511017" Dec 09 10:11:58 crc kubenswrapper[4824]: I1209 10:11:58.446132 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" podStartSLOduration=9.44610507 podStartE2EDuration="9.44610507s" podCreationTimestamp="2025-12-09 10:11:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:11:58.436469424 +0000 UTC m=+1474.770974091" watchObservedRunningTime="2025-12-09 10:11:58.44610507 +0000 UTC m=+1474.780609737" Dec 09 10:11:59 crc kubenswrapper[4824]: I1209 10:11:59.205288 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lhlfr"] Dec 09 10:11:59 crc kubenswrapper[4824]: I1209 10:11:59.673361 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lhlfr" event={"ID":"ad8cfdb4-f7e5-4b30-a1b8-a680b111f788","Type":"ContainerStarted","Data":"68a95d250d8635847290cd9ee6d6506d855bd88c48f7ef78ab67805ad6aa0cb1"} Dec 09 10:12:00 crc kubenswrapper[4824]: I1209 10:12:00.688049 4824 generic.go:334] "Generic (PLEG): container finished" podID="ad8cfdb4-f7e5-4b30-a1b8-a680b111f788" containerID="cc5def6a5a7d96bf7694ed1229224b7871dd16799c2c595a8b42dad978e0ca13" exitCode=0 Dec 09 10:12:00 crc kubenswrapper[4824]: I1209 10:12:00.688139 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lhlfr" event={"ID":"ad8cfdb4-f7e5-4b30-a1b8-a680b111f788","Type":"ContainerDied","Data":"cc5def6a5a7d96bf7694ed1229224b7871dd16799c2c595a8b42dad978e0ca13"} Dec 09 10:12:00 crc kubenswrapper[4824]: I1209 10:12:00.693320 4824 generic.go:334] "Generic (PLEG): container finished" podID="aae4e0f1-0295-4964-8b34-56bdab58a152" containerID="3a153c578db1e11c3e8a67688b76a866e98ce0c6d3d6ff91aca0a32828d69f90" exitCode=0 Dec 09 10:12:00 crc kubenswrapper[4824]: I1209 10:12:00.693395 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ql4lt" event={"ID":"aae4e0f1-0295-4964-8b34-56bdab58a152","Type":"ContainerDied","Data":"3a153c578db1e11c3e8a67688b76a866e98ce0c6d3d6ff91aca0a32828d69f90"} Dec 09 10:12:01 crc kubenswrapper[4824]: I1209 10:12:01.711022 4824 generic.go:334] "Generic (PLEG): container finished" podID="d3d076df-d1e1-473e-9584-79b30cb6d023" containerID="66df44334c000ef2419a325e951be6d72b85973d4249b0d527f40d2ad0d6338b" exitCode=0 Dec 09 10:12:01 crc kubenswrapper[4824]: I1209 10:12:01.711107 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-d4klr" event={"ID":"d3d076df-d1e1-473e-9584-79b30cb6d023","Type":"ContainerDied","Data":"66df44334c000ef2419a325e951be6d72b85973d4249b0d527f40d2ad0d6338b"} Dec 09 10:12:01 crc kubenswrapper[4824]: I1209 10:12:01.713587 4824 generic.go:334] "Generic (PLEG): container finished" podID="6a07abd4-056e-4ea7-962f-60f5a03ea701" containerID="e215988c9459664b8ba3af9c71ac7573e5936c4f048333de2977d582be175243" exitCode=0 Dec 09 10:12:01 crc kubenswrapper[4824]: I1209 10:12:01.713672 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mlcsg" event={"ID":"6a07abd4-056e-4ea7-962f-60f5a03ea701","Type":"ContainerDied","Data":"e215988c9459664b8ba3af9c71ac7573e5936c4f048333de2977d582be175243"} Dec 09 10:12:02 crc kubenswrapper[4824]: I1209 10:12:02.861844 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:12:02 crc kubenswrapper[4824]: I1209 10:12:02.862854 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:12:02 crc kubenswrapper[4824]: I1209 10:12:02.863015 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 10:12:02 crc kubenswrapper[4824]: I1209 10:12:02.864287 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3844f695598277b223637109f124ad02a65103550c5696a49ebd000720021e3c"} pod="openshift-machine-config-operator/machine-config-daemon-dth8x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 10:12:02 crc kubenswrapper[4824]: I1209 10:12:02.864365 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" containerID="cri-o://3844f695598277b223637109f124ad02a65103550c5696a49ebd000720021e3c" gracePeriod=600 Dec 09 10:12:03 crc kubenswrapper[4824]: I1209 10:12:03.739032 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerID="3844f695598277b223637109f124ad02a65103550c5696a49ebd000720021e3c" exitCode=0 Dec 09 10:12:03 crc kubenswrapper[4824]: I1209 10:12:03.739078 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerDied","Data":"3844f695598277b223637109f124ad02a65103550c5696a49ebd000720021e3c"} Dec 09 10:12:03 crc kubenswrapper[4824]: I1209 10:12:03.739380 4824 scope.go:117] "RemoveContainer" containerID="b01bf8ba5eb26e2cf11005bb5cc48bd36d1aa75d59e84d3c71f092c10deef9e7" Dec 09 10:12:05 crc kubenswrapper[4824]: I1209 10:12:05.387412 4824 patch_prober.go:28] interesting pod/thanos-querier-7fc6d7f97-nsq7f container/kube-rbac-proxy-web namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.72:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 10:12:05 crc kubenswrapper[4824]: I1209 10:12:05.387548 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" podUID="d030e6a9-d47a-47bf-9c24-ff0ca58f71e6" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.72:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 10:12:05 crc kubenswrapper[4824]: I1209 10:12:05.708389 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" Dec 09 10:12:05 crc kubenswrapper[4824]: I1209 10:12:05.857394 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-57rtp"] Dec 09 10:12:05 crc kubenswrapper[4824]: I1209 10:12:05.867800 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" podUID="ff1309c3-9703-4b9c-8ec7-e9fdeee87b31" containerName="dnsmasq-dns" containerID="cri-o://75dd9380b91d13bbd28f9a9a2502d7f35cef1d2c681580f2b49e40516fe958ac" gracePeriod=10 Dec 09 10:12:05 crc kubenswrapper[4824]: I1209 10:12:05.869680 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 09 10:12:05 crc kubenswrapper[4824]: I1209 10:12:05.869772 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 09 10:12:05 crc kubenswrapper[4824]: I1209 10:12:05.957014 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 09 10:12:05 crc kubenswrapper[4824]: I1209 10:12:05.977164 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 09 10:12:06 crc kubenswrapper[4824]: I1209 10:12:06.744646 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 09 10:12:06 crc kubenswrapper[4824]: I1209 10:12:06.745812 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 09 10:12:06 crc kubenswrapper[4824]: I1209 10:12:06.836303 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 09 10:12:06 crc kubenswrapper[4824]: I1209 10:12:06.836678 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 09 10:12:06 crc kubenswrapper[4824]: I1209 10:12:06.873991 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 09 10:12:06 crc kubenswrapper[4824]: I1209 10:12:06.885313 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.041991 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" podUID="ff1309c3-9703-4b9c-8ec7-e9fdeee87b31" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.188:5353: connect: connection refused" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.227533 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-d4klr" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.267763 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ql4lt" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.275772 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mlcsg" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.338673 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aae4e0f1-0295-4964-8b34-56bdab58a152-combined-ca-bundle\") pod \"aae4e0f1-0295-4964-8b34-56bdab58a152\" (UID: \"aae4e0f1-0295-4964-8b34-56bdab58a152\") " Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.338731 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3d076df-d1e1-473e-9584-79b30cb6d023-db-sync-config-data\") pod \"d3d076df-d1e1-473e-9584-79b30cb6d023\" (UID: \"d3d076df-d1e1-473e-9584-79b30cb6d023\") " Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.338852 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tz2wp\" (UniqueName: \"kubernetes.io/projected/aae4e0f1-0295-4964-8b34-56bdab58a152-kube-api-access-tz2wp\") pod \"aae4e0f1-0295-4964-8b34-56bdab58a152\" (UID: \"aae4e0f1-0295-4964-8b34-56bdab58a152\") " Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.338883 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mvsf\" (UniqueName: \"kubernetes.io/projected/6a07abd4-056e-4ea7-962f-60f5a03ea701-kube-api-access-5mvsf\") pod \"6a07abd4-056e-4ea7-962f-60f5a03ea701\" (UID: \"6a07abd4-056e-4ea7-962f-60f5a03ea701\") " Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.338918 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-combined-ca-bundle\") pod \"6a07abd4-056e-4ea7-962f-60f5a03ea701\" (UID: \"6a07abd4-056e-4ea7-962f-60f5a03ea701\") " Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.338978 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aae4e0f1-0295-4964-8b34-56bdab58a152-logs\") pod \"aae4e0f1-0295-4964-8b34-56bdab58a152\" (UID: \"aae4e0f1-0295-4964-8b34-56bdab58a152\") " Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.339068 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-config-data\") pod \"6a07abd4-056e-4ea7-962f-60f5a03ea701\" (UID: \"6a07abd4-056e-4ea7-962f-60f5a03ea701\") " Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.339089 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-credential-keys\") pod \"6a07abd4-056e-4ea7-962f-60f5a03ea701\" (UID: \"6a07abd4-056e-4ea7-962f-60f5a03ea701\") " Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.339116 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mh2p2\" (UniqueName: \"kubernetes.io/projected/d3d076df-d1e1-473e-9584-79b30cb6d023-kube-api-access-mh2p2\") pod \"d3d076df-d1e1-473e-9584-79b30cb6d023\" (UID: \"d3d076df-d1e1-473e-9584-79b30cb6d023\") " Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.339172 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3d076df-d1e1-473e-9584-79b30cb6d023-combined-ca-bundle\") pod \"d3d076df-d1e1-473e-9584-79b30cb6d023\" (UID: \"d3d076df-d1e1-473e-9584-79b30cb6d023\") " Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.339194 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-fernet-keys\") pod \"6a07abd4-056e-4ea7-962f-60f5a03ea701\" (UID: \"6a07abd4-056e-4ea7-962f-60f5a03ea701\") " Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.339222 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aae4e0f1-0295-4964-8b34-56bdab58a152-config-data\") pod \"aae4e0f1-0295-4964-8b34-56bdab58a152\" (UID: \"aae4e0f1-0295-4964-8b34-56bdab58a152\") " Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.339266 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-scripts\") pod \"6a07abd4-056e-4ea7-962f-60f5a03ea701\" (UID: \"6a07abd4-056e-4ea7-962f-60f5a03ea701\") " Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.339396 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aae4e0f1-0295-4964-8b34-56bdab58a152-scripts\") pod \"aae4e0f1-0295-4964-8b34-56bdab58a152\" (UID: \"aae4e0f1-0295-4964-8b34-56bdab58a152\") " Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.346924 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-scripts" (OuterVolumeSpecName: "scripts") pod "6a07abd4-056e-4ea7-962f-60f5a03ea701" (UID: "6a07abd4-056e-4ea7-962f-60f5a03ea701"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.358607 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a07abd4-056e-4ea7-962f-60f5a03ea701-kube-api-access-5mvsf" (OuterVolumeSpecName: "kube-api-access-5mvsf") pod "6a07abd4-056e-4ea7-962f-60f5a03ea701" (UID: "6a07abd4-056e-4ea7-962f-60f5a03ea701"). InnerVolumeSpecName "kube-api-access-5mvsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.378974 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3d076df-d1e1-473e-9584-79b30cb6d023-kube-api-access-mh2p2" (OuterVolumeSpecName: "kube-api-access-mh2p2") pod "d3d076df-d1e1-473e-9584-79b30cb6d023" (UID: "d3d076df-d1e1-473e-9584-79b30cb6d023"). InnerVolumeSpecName "kube-api-access-mh2p2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.379802 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6a07abd4-056e-4ea7-962f-60f5a03ea701" (UID: "6a07abd4-056e-4ea7-962f-60f5a03ea701"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.379866 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "6a07abd4-056e-4ea7-962f-60f5a03ea701" (UID: "6a07abd4-056e-4ea7-962f-60f5a03ea701"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.379837 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aae4e0f1-0295-4964-8b34-56bdab58a152-logs" (OuterVolumeSpecName: "logs") pod "aae4e0f1-0295-4964-8b34-56bdab58a152" (UID: "aae4e0f1-0295-4964-8b34-56bdab58a152"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.380564 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3d076df-d1e1-473e-9584-79b30cb6d023-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d3d076df-d1e1-473e-9584-79b30cb6d023" (UID: "d3d076df-d1e1-473e-9584-79b30cb6d023"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.382343 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aae4e0f1-0295-4964-8b34-56bdab58a152-kube-api-access-tz2wp" (OuterVolumeSpecName: "kube-api-access-tz2wp") pod "aae4e0f1-0295-4964-8b34-56bdab58a152" (UID: "aae4e0f1-0295-4964-8b34-56bdab58a152"). InnerVolumeSpecName "kube-api-access-tz2wp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.386732 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aae4e0f1-0295-4964-8b34-56bdab58a152-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aae4e0f1-0295-4964-8b34-56bdab58a152" (UID: "aae4e0f1-0295-4964-8b34-56bdab58a152"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.389263 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aae4e0f1-0295-4964-8b34-56bdab58a152-scripts" (OuterVolumeSpecName: "scripts") pod "aae4e0f1-0295-4964-8b34-56bdab58a152" (UID: "aae4e0f1-0295-4964-8b34-56bdab58a152"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.410735 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aae4e0f1-0295-4964-8b34-56bdab58a152-config-data" (OuterVolumeSpecName: "config-data") pod "aae4e0f1-0295-4964-8b34-56bdab58a152" (UID: "aae4e0f1-0295-4964-8b34-56bdab58a152"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.416676 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3d076df-d1e1-473e-9584-79b30cb6d023-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3d076df-d1e1-473e-9584-79b30cb6d023" (UID: "d3d076df-d1e1-473e-9584-79b30cb6d023"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.451077 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tz2wp\" (UniqueName: \"kubernetes.io/projected/aae4e0f1-0295-4964-8b34-56bdab58a152-kube-api-access-tz2wp\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.451122 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mvsf\" (UniqueName: \"kubernetes.io/projected/6a07abd4-056e-4ea7-962f-60f5a03ea701-kube-api-access-5mvsf\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.451135 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aae4e0f1-0295-4964-8b34-56bdab58a152-logs\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.451148 4824 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.451161 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mh2p2\" (UniqueName: \"kubernetes.io/projected/d3d076df-d1e1-473e-9584-79b30cb6d023-kube-api-access-mh2p2\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.451172 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3d076df-d1e1-473e-9584-79b30cb6d023-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.451184 4824 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.451195 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aae4e0f1-0295-4964-8b34-56bdab58a152-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.451206 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.451216 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aae4e0f1-0295-4964-8b34-56bdab58a152-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.451227 4824 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3d076df-d1e1-473e-9584-79b30cb6d023-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.451243 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aae4e0f1-0295-4964-8b34-56bdab58a152-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.460252 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-config-data" (OuterVolumeSpecName: "config-data") pod "6a07abd4-056e-4ea7-962f-60f5a03ea701" (UID: "6a07abd4-056e-4ea7-962f-60f5a03ea701"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.534632 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a07abd4-056e-4ea7-962f-60f5a03ea701" (UID: "6a07abd4-056e-4ea7-962f-60f5a03ea701"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.557614 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.557653 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a07abd4-056e-4ea7-962f-60f5a03ea701-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.607337 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.769131 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-config\") pod \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\" (UID: \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\") " Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.769427 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-ovsdbserver-sb\") pod \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\" (UID: \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\") " Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.769476 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-dns-svc\") pod \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\" (UID: \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\") " Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.769923 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-ovsdbserver-nb\") pod \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\" (UID: \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\") " Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.769957 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-dns-swift-storage-0\") pod \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\" (UID: \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\") " Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.770059 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4qnv\" (UniqueName: \"kubernetes.io/projected/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-kube-api-access-t4qnv\") pod \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\" (UID: \"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31\") " Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.791044 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mlcsg" event={"ID":"6a07abd4-056e-4ea7-962f-60f5a03ea701","Type":"ContainerDied","Data":"a4b5f0cd195efa7bf674fc7266a9484f3d9672b8cc14cc0b462b43fd6440895a"} Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.791099 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4b5f0cd195efa7bf674fc7266a9484f3d9672b8cc14cc0b462b43fd6440895a" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.791171 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mlcsg" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.798493 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-vfnsw" event={"ID":"405d0a5e-fd62-4146-bfb9-96ff6c077836","Type":"ContainerStarted","Data":"40ff06ab3f0f7bbf01219bb470c28376ccf91ab46f772e08dc41ca02ca67a752"} Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.802992 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-kube-api-access-t4qnv" (OuterVolumeSpecName: "kube-api-access-t4qnv") pod "ff1309c3-9703-4b9c-8ec7-e9fdeee87b31" (UID: "ff1309c3-9703-4b9c-8ec7-e9fdeee87b31"). InnerVolumeSpecName "kube-api-access-t4qnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.806146 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ql4lt" event={"ID":"aae4e0f1-0295-4964-8b34-56bdab58a152","Type":"ContainerDied","Data":"445f3f2271e4d78cb53e9d0ce785081f27feb5a77762defdfb2a15209f6d44a3"} Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.806203 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="445f3f2271e4d78cb53e9d0ce785081f27feb5a77762defdfb2a15209f6d44a3" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.806294 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ql4lt" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.810992 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-d4klr" event={"ID":"d3d076df-d1e1-473e-9584-79b30cb6d023","Type":"ContainerDied","Data":"8857606b103f540643d96ffaa2f25674c464039ebffbf499c32645926ab0d0f8"} Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.811044 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8857606b103f540643d96ffaa2f25674c464039ebffbf499c32645926ab0d0f8" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.811120 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-d4klr" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.818350 4824 generic.go:334] "Generic (PLEG): container finished" podID="ff1309c3-9703-4b9c-8ec7-e9fdeee87b31" containerID="75dd9380b91d13bbd28f9a9a2502d7f35cef1d2c681580f2b49e40516fe958ac" exitCode=0 Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.818473 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" event={"ID":"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31","Type":"ContainerDied","Data":"75dd9380b91d13bbd28f9a9a2502d7f35cef1d2c681580f2b49e40516fe958ac"} Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.818515 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" event={"ID":"ff1309c3-9703-4b9c-8ec7-e9fdeee87b31","Type":"ContainerDied","Data":"61b4effa6402fb58c433e87b2da3cb8f5ec47876048837b2193ab8e4b7737e24"} Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.818538 4824 scope.go:117] "RemoveContainer" containerID="75dd9380b91d13bbd28f9a9a2502d7f35cef1d2c681580f2b49e40516fe958ac" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.818741 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-57rtp" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.832490 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-vfnsw" podStartSLOduration=11.291466618 podStartE2EDuration="53.832461293s" podCreationTimestamp="2025-12-09 10:11:14 +0000 UTC" firstStartedPulling="2025-12-09 10:11:17.842967486 +0000 UTC m=+1434.177472153" lastFinishedPulling="2025-12-09 10:12:00.383962161 +0000 UTC m=+1476.718466828" observedRunningTime="2025-12-09 10:12:07.825562643 +0000 UTC m=+1484.160067320" watchObservedRunningTime="2025-12-09 10:12:07.832461293 +0000 UTC m=+1484.166965960" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.836617 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.836656 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.874218 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4qnv\" (UniqueName: \"kubernetes.io/projected/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-kube-api-access-t4qnv\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.884122 4824 scope.go:117] "RemoveContainer" containerID="cc96b5d9ac0d22837e051873be369805cd0878b0bb2394919a4662ab448cb53b" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.986332 4824 scope.go:117] "RemoveContainer" containerID="75dd9380b91d13bbd28f9a9a2502d7f35cef1d2c681580f2b49e40516fe958ac" Dec 09 10:12:07 crc kubenswrapper[4824]: E1209 10:12:07.988874 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75dd9380b91d13bbd28f9a9a2502d7f35cef1d2c681580f2b49e40516fe958ac\": container with ID starting with 75dd9380b91d13bbd28f9a9a2502d7f35cef1d2c681580f2b49e40516fe958ac not found: ID does not exist" containerID="75dd9380b91d13bbd28f9a9a2502d7f35cef1d2c681580f2b49e40516fe958ac" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.988959 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75dd9380b91d13bbd28f9a9a2502d7f35cef1d2c681580f2b49e40516fe958ac"} err="failed to get container status \"75dd9380b91d13bbd28f9a9a2502d7f35cef1d2c681580f2b49e40516fe958ac\": rpc error: code = NotFound desc = could not find container \"75dd9380b91d13bbd28f9a9a2502d7f35cef1d2c681580f2b49e40516fe958ac\": container with ID starting with 75dd9380b91d13bbd28f9a9a2502d7f35cef1d2c681580f2b49e40516fe958ac not found: ID does not exist" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.988980 4824 scope.go:117] "RemoveContainer" containerID="cc96b5d9ac0d22837e051873be369805cd0878b0bb2394919a4662ab448cb53b" Dec 09 10:12:07 crc kubenswrapper[4824]: E1209 10:12:07.989403 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc96b5d9ac0d22837e051873be369805cd0878b0bb2394919a4662ab448cb53b\": container with ID starting with cc96b5d9ac0d22837e051873be369805cd0878b0bb2394919a4662ab448cb53b not found: ID does not exist" containerID="cc96b5d9ac0d22837e051873be369805cd0878b0bb2394919a4662ab448cb53b" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.989427 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc96b5d9ac0d22837e051873be369805cd0878b0bb2394919a4662ab448cb53b"} err="failed to get container status \"cc96b5d9ac0d22837e051873be369805cd0878b0bb2394919a4662ab448cb53b\": rpc error: code = NotFound desc = could not find container \"cc96b5d9ac0d22837e051873be369805cd0878b0bb2394919a4662ab448cb53b\": container with ID starting with cc96b5d9ac0d22837e051873be369805cd0878b0bb2394919a4662ab448cb53b not found: ID does not exist" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.995729 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ff1309c3-9703-4b9c-8ec7-e9fdeee87b31" (UID: "ff1309c3-9703-4b9c-8ec7-e9fdeee87b31"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:12:07 crc kubenswrapper[4824]: I1209 10:12:07.994684 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ff1309c3-9703-4b9c-8ec7-e9fdeee87b31" (UID: "ff1309c3-9703-4b9c-8ec7-e9fdeee87b31"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.027768 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-config" (OuterVolumeSpecName: "config") pod "ff1309c3-9703-4b9c-8ec7-e9fdeee87b31" (UID: "ff1309c3-9703-4b9c-8ec7-e9fdeee87b31"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.028215 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ff1309c3-9703-4b9c-8ec7-e9fdeee87b31" (UID: "ff1309c3-9703-4b9c-8ec7-e9fdeee87b31"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.041304 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ff1309c3-9703-4b9c-8ec7-e9fdeee87b31" (UID: "ff1309c3-9703-4b9c-8ec7-e9fdeee87b31"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.079439 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.079502 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.079518 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.079533 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.079546 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.226877 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-57rtp"] Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.247991 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-57rtp"] Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.550683 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-55b96fff66-brzz9"] Dec 09 10:12:08 crc kubenswrapper[4824]: E1209 10:12:08.551696 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a07abd4-056e-4ea7-962f-60f5a03ea701" containerName="keystone-bootstrap" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.551717 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a07abd4-056e-4ea7-962f-60f5a03ea701" containerName="keystone-bootstrap" Dec 09 10:12:08 crc kubenswrapper[4824]: E1209 10:12:08.551739 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff1309c3-9703-4b9c-8ec7-e9fdeee87b31" containerName="init" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.551746 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff1309c3-9703-4b9c-8ec7-e9fdeee87b31" containerName="init" Dec 09 10:12:08 crc kubenswrapper[4824]: E1209 10:12:08.551791 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae4e0f1-0295-4964-8b34-56bdab58a152" containerName="placement-db-sync" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.551798 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae4e0f1-0295-4964-8b34-56bdab58a152" containerName="placement-db-sync" Dec 09 10:12:08 crc kubenswrapper[4824]: E1209 10:12:08.551811 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3d076df-d1e1-473e-9584-79b30cb6d023" containerName="barbican-db-sync" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.551818 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3d076df-d1e1-473e-9584-79b30cb6d023" containerName="barbican-db-sync" Dec 09 10:12:08 crc kubenswrapper[4824]: E1209 10:12:08.551836 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff1309c3-9703-4b9c-8ec7-e9fdeee87b31" containerName="dnsmasq-dns" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.551842 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff1309c3-9703-4b9c-8ec7-e9fdeee87b31" containerName="dnsmasq-dns" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.552081 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a07abd4-056e-4ea7-962f-60f5a03ea701" containerName="keystone-bootstrap" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.552109 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff1309c3-9703-4b9c-8ec7-e9fdeee87b31" containerName="dnsmasq-dns" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.552122 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="aae4e0f1-0295-4964-8b34-56bdab58a152" containerName="placement-db-sync" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.552135 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3d076df-d1e1-473e-9584-79b30cb6d023" containerName="barbican-db-sync" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.553508 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-55b96fff66-brzz9" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.562476 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.562692 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.562967 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-gdb54" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.572373 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-77bcf58d55-bshs5"] Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.575389 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-77bcf58d55-bshs5" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.597140 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-77bcf58d55-bshs5"] Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.621247 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.668966 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7c7958c67b-wr8zx"] Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.671908 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.680414 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.680638 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.680819 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-9snwp" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.681031 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.681181 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.716565 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d72ff513-d8d9-4191-92d4-7824fca18f32-config-data-custom\") pod \"barbican-keystone-listener-55b96fff66-brzz9\" (UID: \"d72ff513-d8d9-4191-92d4-7824fca18f32\") " pod="openstack/barbican-keystone-listener-55b96fff66-brzz9" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.716666 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d72ff513-d8d9-4191-92d4-7824fca18f32-combined-ca-bundle\") pod \"barbican-keystone-listener-55b96fff66-brzz9\" (UID: \"d72ff513-d8d9-4191-92d4-7824fca18f32\") " pod="openstack/barbican-keystone-listener-55b96fff66-brzz9" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.716701 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/555ef974-afb9-4455-8a15-c728200be8a2-combined-ca-bundle\") pod \"barbican-worker-77bcf58d55-bshs5\" (UID: \"555ef974-afb9-4455-8a15-c728200be8a2\") " pod="openstack/barbican-worker-77bcf58d55-bshs5" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.716736 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/555ef974-afb9-4455-8a15-c728200be8a2-logs\") pod \"barbican-worker-77bcf58d55-bshs5\" (UID: \"555ef974-afb9-4455-8a15-c728200be8a2\") " pod="openstack/barbican-worker-77bcf58d55-bshs5" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.716762 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6htvv\" (UniqueName: \"kubernetes.io/projected/555ef974-afb9-4455-8a15-c728200be8a2-kube-api-access-6htvv\") pod \"barbican-worker-77bcf58d55-bshs5\" (UID: \"555ef974-afb9-4455-8a15-c728200be8a2\") " pod="openstack/barbican-worker-77bcf58d55-bshs5" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.716814 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d72ff513-d8d9-4191-92d4-7824fca18f32-logs\") pod \"barbican-keystone-listener-55b96fff66-brzz9\" (UID: \"d72ff513-d8d9-4191-92d4-7824fca18f32\") " pod="openstack/barbican-keystone-listener-55b96fff66-brzz9" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.716839 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/555ef974-afb9-4455-8a15-c728200be8a2-config-data-custom\") pod \"barbican-worker-77bcf58d55-bshs5\" (UID: \"555ef974-afb9-4455-8a15-c728200be8a2\") " pod="openstack/barbican-worker-77bcf58d55-bshs5" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.716903 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/555ef974-afb9-4455-8a15-c728200be8a2-config-data\") pod \"barbican-worker-77bcf58d55-bshs5\" (UID: \"555ef974-afb9-4455-8a15-c728200be8a2\") " pod="openstack/barbican-worker-77bcf58d55-bshs5" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.717002 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47rkq\" (UniqueName: \"kubernetes.io/projected/d72ff513-d8d9-4191-92d4-7824fca18f32-kube-api-access-47rkq\") pod \"barbican-keystone-listener-55b96fff66-brzz9\" (UID: \"d72ff513-d8d9-4191-92d4-7824fca18f32\") " pod="openstack/barbican-keystone-listener-55b96fff66-brzz9" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.717069 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d72ff513-d8d9-4191-92d4-7824fca18f32-config-data\") pod \"barbican-keystone-listener-55b96fff66-brzz9\" (UID: \"d72ff513-d8d9-4191-92d4-7824fca18f32\") " pod="openstack/barbican-keystone-listener-55b96fff66-brzz9" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.740641 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-9c7879457-2fmzw"] Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.743411 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.753413 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.753776 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.754151 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.754400 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ggbf4" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.754547 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.755014 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.782146 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-55b96fff66-brzz9"] Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.803569 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7c7958c67b-wr8zx"] Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.828907 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/555ef974-afb9-4455-8a15-c728200be8a2-logs\") pod \"barbican-worker-77bcf58d55-bshs5\" (UID: \"555ef974-afb9-4455-8a15-c728200be8a2\") " pod="openstack/barbican-worker-77bcf58d55-bshs5" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.828967 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9dcb334f-65bf-4f84-88a0-24ca56b4a8c7-scripts\") pod \"keystone-9c7879457-2fmzw\" (UID: \"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7\") " pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.828997 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6htvv\" (UniqueName: \"kubernetes.io/projected/555ef974-afb9-4455-8a15-c728200be8a2-kube-api-access-6htvv\") pod \"barbican-worker-77bcf58d55-bshs5\" (UID: \"555ef974-afb9-4455-8a15-c728200be8a2\") " pod="openstack/barbican-worker-77bcf58d55-bshs5" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.829038 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d72ff513-d8d9-4191-92d4-7824fca18f32-logs\") pod \"barbican-keystone-listener-55b96fff66-brzz9\" (UID: \"d72ff513-d8d9-4191-92d4-7824fca18f32\") " pod="openstack/barbican-keystone-listener-55b96fff66-brzz9" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.829062 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/555ef974-afb9-4455-8a15-c728200be8a2-config-data-custom\") pod \"barbican-worker-77bcf58d55-bshs5\" (UID: \"555ef974-afb9-4455-8a15-c728200be8a2\") " pod="openstack/barbican-worker-77bcf58d55-bshs5" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.829107 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/555ef974-afb9-4455-8a15-c728200be8a2-config-data\") pod \"barbican-worker-77bcf58d55-bshs5\" (UID: \"555ef974-afb9-4455-8a15-c728200be8a2\") " pod="openstack/barbican-worker-77bcf58d55-bshs5" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.829136 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dcb334f-65bf-4f84-88a0-24ca56b4a8c7-public-tls-certs\") pod \"keystone-9c7879457-2fmzw\" (UID: \"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7\") " pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.829160 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38c32e83-50e7-4f8b-8bbf-26adebff1b82-public-tls-certs\") pod \"placement-7c7958c67b-wr8zx\" (UID: \"38c32e83-50e7-4f8b-8bbf-26adebff1b82\") " pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.829185 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/38c32e83-50e7-4f8b-8bbf-26adebff1b82-internal-tls-certs\") pod \"placement-7c7958c67b-wr8zx\" (UID: \"38c32e83-50e7-4f8b-8bbf-26adebff1b82\") " pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.829198 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dcb334f-65bf-4f84-88a0-24ca56b4a8c7-internal-tls-certs\") pod \"keystone-9c7879457-2fmzw\" (UID: \"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7\") " pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.829226 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38c32e83-50e7-4f8b-8bbf-26adebff1b82-scripts\") pod \"placement-7c7958c67b-wr8zx\" (UID: \"38c32e83-50e7-4f8b-8bbf-26adebff1b82\") " pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.829249 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38c32e83-50e7-4f8b-8bbf-26adebff1b82-logs\") pod \"placement-7c7958c67b-wr8zx\" (UID: \"38c32e83-50e7-4f8b-8bbf-26adebff1b82\") " pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.829275 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47rkq\" (UniqueName: \"kubernetes.io/projected/d72ff513-d8d9-4191-92d4-7824fca18f32-kube-api-access-47rkq\") pod \"barbican-keystone-listener-55b96fff66-brzz9\" (UID: \"d72ff513-d8d9-4191-92d4-7824fca18f32\") " pod="openstack/barbican-keystone-listener-55b96fff66-brzz9" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.829296 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxbrt\" (UniqueName: \"kubernetes.io/projected/9dcb334f-65bf-4f84-88a0-24ca56b4a8c7-kube-api-access-cxbrt\") pod \"keystone-9c7879457-2fmzw\" (UID: \"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7\") " pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.829336 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k52wn\" (UniqueName: \"kubernetes.io/projected/38c32e83-50e7-4f8b-8bbf-26adebff1b82-kube-api-access-k52wn\") pod \"placement-7c7958c67b-wr8zx\" (UID: \"38c32e83-50e7-4f8b-8bbf-26adebff1b82\") " pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.829361 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d72ff513-d8d9-4191-92d4-7824fca18f32-config-data\") pod \"barbican-keystone-listener-55b96fff66-brzz9\" (UID: \"d72ff513-d8d9-4191-92d4-7824fca18f32\") " pod="openstack/barbican-keystone-listener-55b96fff66-brzz9" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.829382 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d72ff513-d8d9-4191-92d4-7824fca18f32-config-data-custom\") pod \"barbican-keystone-listener-55b96fff66-brzz9\" (UID: \"d72ff513-d8d9-4191-92d4-7824fca18f32\") " pod="openstack/barbican-keystone-listener-55b96fff66-brzz9" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.829400 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38c32e83-50e7-4f8b-8bbf-26adebff1b82-combined-ca-bundle\") pod \"placement-7c7958c67b-wr8zx\" (UID: \"38c32e83-50e7-4f8b-8bbf-26adebff1b82\") " pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.829430 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9dcb334f-65bf-4f84-88a0-24ca56b4a8c7-credential-keys\") pod \"keystone-9c7879457-2fmzw\" (UID: \"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7\") " pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.829447 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38c32e83-50e7-4f8b-8bbf-26adebff1b82-config-data\") pod \"placement-7c7958c67b-wr8zx\" (UID: \"38c32e83-50e7-4f8b-8bbf-26adebff1b82\") " pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.829478 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9dcb334f-65bf-4f84-88a0-24ca56b4a8c7-fernet-keys\") pod \"keystone-9c7879457-2fmzw\" (UID: \"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7\") " pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.829516 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d72ff513-d8d9-4191-92d4-7824fca18f32-combined-ca-bundle\") pod \"barbican-keystone-listener-55b96fff66-brzz9\" (UID: \"d72ff513-d8d9-4191-92d4-7824fca18f32\") " pod="openstack/barbican-keystone-listener-55b96fff66-brzz9" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.829543 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dcb334f-65bf-4f84-88a0-24ca56b4a8c7-config-data\") pod \"keystone-9c7879457-2fmzw\" (UID: \"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7\") " pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.829562 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/555ef974-afb9-4455-8a15-c728200be8a2-combined-ca-bundle\") pod \"barbican-worker-77bcf58d55-bshs5\" (UID: \"555ef974-afb9-4455-8a15-c728200be8a2\") " pod="openstack/barbican-worker-77bcf58d55-bshs5" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.829578 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dcb334f-65bf-4f84-88a0-24ca56b4a8c7-combined-ca-bundle\") pod \"keystone-9c7879457-2fmzw\" (UID: \"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7\") " pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.830046 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/555ef974-afb9-4455-8a15-c728200be8a2-logs\") pod \"barbican-worker-77bcf58d55-bshs5\" (UID: \"555ef974-afb9-4455-8a15-c728200be8a2\") " pod="openstack/barbican-worker-77bcf58d55-bshs5" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.830578 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d72ff513-d8d9-4191-92d4-7824fca18f32-logs\") pod \"barbican-keystone-listener-55b96fff66-brzz9\" (UID: \"d72ff513-d8d9-4191-92d4-7824fca18f32\") " pod="openstack/barbican-keystone-listener-55b96fff66-brzz9" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.841957 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d72ff513-d8d9-4191-92d4-7824fca18f32-config-data-custom\") pod \"barbican-keystone-listener-55b96fff66-brzz9\" (UID: \"d72ff513-d8d9-4191-92d4-7824fca18f32\") " pod="openstack/barbican-keystone-listener-55b96fff66-brzz9" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.842315 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d72ff513-d8d9-4191-92d4-7824fca18f32-config-data\") pod \"barbican-keystone-listener-55b96fff66-brzz9\" (UID: \"d72ff513-d8d9-4191-92d4-7824fca18f32\") " pod="openstack/barbican-keystone-listener-55b96fff66-brzz9" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.855942 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d72ff513-d8d9-4191-92d4-7824fca18f32-combined-ca-bundle\") pod \"barbican-keystone-listener-55b96fff66-brzz9\" (UID: \"d72ff513-d8d9-4191-92d4-7824fca18f32\") " pod="openstack/barbican-keystone-listener-55b96fff66-brzz9" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.861038 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47rkq\" (UniqueName: \"kubernetes.io/projected/d72ff513-d8d9-4191-92d4-7824fca18f32-kube-api-access-47rkq\") pod \"barbican-keystone-listener-55b96fff66-brzz9\" (UID: \"d72ff513-d8d9-4191-92d4-7824fca18f32\") " pod="openstack/barbican-keystone-listener-55b96fff66-brzz9" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.862219 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/555ef974-afb9-4455-8a15-c728200be8a2-combined-ca-bundle\") pod \"barbican-worker-77bcf58d55-bshs5\" (UID: \"555ef974-afb9-4455-8a15-c728200be8a2\") " pod="openstack/barbican-worker-77bcf58d55-bshs5" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.865000 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/555ef974-afb9-4455-8a15-c728200be8a2-config-data\") pod \"barbican-worker-77bcf58d55-bshs5\" (UID: \"555ef974-afb9-4455-8a15-c728200be8a2\") " pod="openstack/barbican-worker-77bcf58d55-bshs5" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.868453 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-9c7879457-2fmzw"] Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.871495 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/555ef974-afb9-4455-8a15-c728200be8a2-config-data-custom\") pod \"barbican-worker-77bcf58d55-bshs5\" (UID: \"555ef974-afb9-4455-8a15-c728200be8a2\") " pod="openstack/barbican-worker-77bcf58d55-bshs5" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.886557 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6htvv\" (UniqueName: \"kubernetes.io/projected/555ef974-afb9-4455-8a15-c728200be8a2-kube-api-access-6htvv\") pod \"barbican-worker-77bcf58d55-bshs5\" (UID: \"555ef974-afb9-4455-8a15-c728200be8a2\") " pod="openstack/barbican-worker-77bcf58d55-bshs5" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.887387 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"232c4238-3f59-4e05-ac75-8746ab0a0069","Type":"ContainerStarted","Data":"332f494a5454b4367146a6f5595115254e0066e1d68d691290d0464175c57efe"} Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.932188 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-55b96fff66-brzz9" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.938071 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dcb334f-65bf-4f84-88a0-24ca56b4a8c7-public-tls-certs\") pod \"keystone-9c7879457-2fmzw\" (UID: \"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7\") " pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.938148 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38c32e83-50e7-4f8b-8bbf-26adebff1b82-public-tls-certs\") pod \"placement-7c7958c67b-wr8zx\" (UID: \"38c32e83-50e7-4f8b-8bbf-26adebff1b82\") " pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.938186 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/38c32e83-50e7-4f8b-8bbf-26adebff1b82-internal-tls-certs\") pod \"placement-7c7958c67b-wr8zx\" (UID: \"38c32e83-50e7-4f8b-8bbf-26adebff1b82\") " pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.938204 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dcb334f-65bf-4f84-88a0-24ca56b4a8c7-internal-tls-certs\") pod \"keystone-9c7879457-2fmzw\" (UID: \"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7\") " pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.938272 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38c32e83-50e7-4f8b-8bbf-26adebff1b82-scripts\") pod \"placement-7c7958c67b-wr8zx\" (UID: \"38c32e83-50e7-4f8b-8bbf-26adebff1b82\") " pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.938315 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38c32e83-50e7-4f8b-8bbf-26adebff1b82-logs\") pod \"placement-7c7958c67b-wr8zx\" (UID: \"38c32e83-50e7-4f8b-8bbf-26adebff1b82\") " pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.938374 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxbrt\" (UniqueName: \"kubernetes.io/projected/9dcb334f-65bf-4f84-88a0-24ca56b4a8c7-kube-api-access-cxbrt\") pod \"keystone-9c7879457-2fmzw\" (UID: \"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7\") " pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.938456 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k52wn\" (UniqueName: \"kubernetes.io/projected/38c32e83-50e7-4f8b-8bbf-26adebff1b82-kube-api-access-k52wn\") pod \"placement-7c7958c67b-wr8zx\" (UID: \"38c32e83-50e7-4f8b-8bbf-26adebff1b82\") " pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.938515 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38c32e83-50e7-4f8b-8bbf-26adebff1b82-combined-ca-bundle\") pod \"placement-7c7958c67b-wr8zx\" (UID: \"38c32e83-50e7-4f8b-8bbf-26adebff1b82\") " pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.938581 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9dcb334f-65bf-4f84-88a0-24ca56b4a8c7-credential-keys\") pod \"keystone-9c7879457-2fmzw\" (UID: \"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7\") " pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.938607 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38c32e83-50e7-4f8b-8bbf-26adebff1b82-config-data\") pod \"placement-7c7958c67b-wr8zx\" (UID: \"38c32e83-50e7-4f8b-8bbf-26adebff1b82\") " pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.938683 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9dcb334f-65bf-4f84-88a0-24ca56b4a8c7-fernet-keys\") pod \"keystone-9c7879457-2fmzw\" (UID: \"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7\") " pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.938768 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dcb334f-65bf-4f84-88a0-24ca56b4a8c7-config-data\") pod \"keystone-9c7879457-2fmzw\" (UID: \"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7\") " pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.938824 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dcb334f-65bf-4f84-88a0-24ca56b4a8c7-combined-ca-bundle\") pod \"keystone-9c7879457-2fmzw\" (UID: \"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7\") " pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.938901 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9dcb334f-65bf-4f84-88a0-24ca56b4a8c7-scripts\") pod \"keystone-9c7879457-2fmzw\" (UID: \"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7\") " pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.946252 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9dcb334f-65bf-4f84-88a0-24ca56b4a8c7-fernet-keys\") pod \"keystone-9c7879457-2fmzw\" (UID: \"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7\") " pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.945840 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38c32e83-50e7-4f8b-8bbf-26adebff1b82-logs\") pod \"placement-7c7958c67b-wr8zx\" (UID: \"38c32e83-50e7-4f8b-8bbf-26adebff1b82\") " pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.954193 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerStarted","Data":"7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1"} Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.964203 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lhlfr" event={"ID":"ad8cfdb4-f7e5-4b30-a1b8-a680b111f788","Type":"ContainerStarted","Data":"c0650433deec5abea39f2881e472191bd159aadb67a9c737ba5bd6be7c0a63c9"} Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.967727 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-jghjq" event={"ID":"dbceafb7-272a-45fc-9b56-df631986f90a","Type":"ContainerStarted","Data":"751410b43bd034373b18f9792b5d576ad6b6aa7d073a4a19569c94413f9c45c0"} Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.969610 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38c32e83-50e7-4f8b-8bbf-26adebff1b82-scripts\") pod \"placement-7c7958c67b-wr8zx\" (UID: \"38c32e83-50e7-4f8b-8bbf-26adebff1b82\") " pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.973532 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dcb334f-65bf-4f84-88a0-24ca56b4a8c7-internal-tls-certs\") pod \"keystone-9c7879457-2fmzw\" (UID: \"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7\") " pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.973547 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38c32e83-50e7-4f8b-8bbf-26adebff1b82-config-data\") pod \"placement-7c7958c67b-wr8zx\" (UID: \"38c32e83-50e7-4f8b-8bbf-26adebff1b82\") " pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.973963 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k52wn\" (UniqueName: \"kubernetes.io/projected/38c32e83-50e7-4f8b-8bbf-26adebff1b82-kube-api-access-k52wn\") pod \"placement-7c7958c67b-wr8zx\" (UID: \"38c32e83-50e7-4f8b-8bbf-26adebff1b82\") " pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.981613 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9dcb334f-65bf-4f84-88a0-24ca56b4a8c7-credential-keys\") pod \"keystone-9c7879457-2fmzw\" (UID: \"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7\") " pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.982285 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dcb334f-65bf-4f84-88a0-24ca56b4a8c7-config-data\") pod \"keystone-9c7879457-2fmzw\" (UID: \"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7\") " pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.985946 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/38c32e83-50e7-4f8b-8bbf-26adebff1b82-internal-tls-certs\") pod \"placement-7c7958c67b-wr8zx\" (UID: \"38c32e83-50e7-4f8b-8bbf-26adebff1b82\") " pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.986508 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dcb334f-65bf-4f84-88a0-24ca56b4a8c7-public-tls-certs\") pod \"keystone-9c7879457-2fmzw\" (UID: \"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7\") " pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.993181 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-77bcf58d55-bshs5" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.994191 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9dcb334f-65bf-4f84-88a0-24ca56b4a8c7-scripts\") pod \"keystone-9c7879457-2fmzw\" (UID: \"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7\") " pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:08 crc kubenswrapper[4824]: I1209 10:12:08.999912 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxbrt\" (UniqueName: \"kubernetes.io/projected/9dcb334f-65bf-4f84-88a0-24ca56b4a8c7-kube-api-access-cxbrt\") pod \"keystone-9c7879457-2fmzw\" (UID: \"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7\") " pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.000026 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-2952m"] Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.003024 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-2952m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.012178 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-2952m"] Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.014340 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38c32e83-50e7-4f8b-8bbf-26adebff1b82-combined-ca-bundle\") pod \"placement-7c7958c67b-wr8zx\" (UID: \"38c32e83-50e7-4f8b-8bbf-26adebff1b82\") " pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.019601 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dcb334f-65bf-4f84-88a0-24ca56b4a8c7-combined-ca-bundle\") pod \"keystone-9c7879457-2fmzw\" (UID: \"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7\") " pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.022108 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38c32e83-50e7-4f8b-8bbf-26adebff1b82-public-tls-certs\") pod \"placement-7c7958c67b-wr8zx\" (UID: \"38c32e83-50e7-4f8b-8bbf-26adebff1b82\") " pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.029299 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.055863 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5f9b44b4db-xxc6m"] Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.065680 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f9b44b4db-xxc6m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.079542 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.081104 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.088979 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5f9b44b4db-xxc6m"] Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.142204 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-jghjq" podStartSLOduration=5.73945749 podStartE2EDuration="55.142184951s" podCreationTimestamp="2025-12-09 10:11:14 +0000 UTC" firstStartedPulling="2025-12-09 10:11:17.868968203 +0000 UTC m=+1434.203472870" lastFinishedPulling="2025-12-09 10:12:07.271695664 +0000 UTC m=+1483.606200331" observedRunningTime="2025-12-09 10:12:09.133278717 +0000 UTC m=+1485.467783394" watchObservedRunningTime="2025-12-09 10:12:09.142184951 +0000 UTC m=+1485.476689618" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.143668 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f8a567c-2577-4c66-a75e-e43f7cd66000-config-data\") pod \"barbican-api-5f9b44b4db-xxc6m\" (UID: \"9f8a567c-2577-4c66-a75e-e43f7cd66000\") " pod="openstack/barbican-api-5f9b44b4db-xxc6m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.143859 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-2952m\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " pod="openstack/dnsmasq-dns-85ff748b95-2952m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.143893 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-2952m\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " pod="openstack/dnsmasq-dns-85ff748b95-2952m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.143915 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f8a567c-2577-4c66-a75e-e43f7cd66000-config-data-custom\") pod \"barbican-api-5f9b44b4db-xxc6m\" (UID: \"9f8a567c-2577-4c66-a75e-e43f7cd66000\") " pod="openstack/barbican-api-5f9b44b4db-xxc6m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.143932 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-2952m\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " pod="openstack/dnsmasq-dns-85ff748b95-2952m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.143997 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f8a567c-2577-4c66-a75e-e43f7cd66000-logs\") pod \"barbican-api-5f9b44b4db-xxc6m\" (UID: \"9f8a567c-2577-4c66-a75e-e43f7cd66000\") " pod="openstack/barbican-api-5f9b44b4db-xxc6m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.144021 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-config\") pod \"dnsmasq-dns-85ff748b95-2952m\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " pod="openstack/dnsmasq-dns-85ff748b95-2952m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.145290 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f8a567c-2577-4c66-a75e-e43f7cd66000-combined-ca-bundle\") pod \"barbican-api-5f9b44b4db-xxc6m\" (UID: \"9f8a567c-2577-4c66-a75e-e43f7cd66000\") " pod="openstack/barbican-api-5f9b44b4db-xxc6m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.145507 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp4c2\" (UniqueName: \"kubernetes.io/projected/9f8a567c-2577-4c66-a75e-e43f7cd66000-kube-api-access-jp4c2\") pod \"barbican-api-5f9b44b4db-xxc6m\" (UID: \"9f8a567c-2577-4c66-a75e-e43f7cd66000\") " pod="openstack/barbican-api-5f9b44b4db-xxc6m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.145800 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-dns-svc\") pod \"dnsmasq-dns-85ff748b95-2952m\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " pod="openstack/dnsmasq-dns-85ff748b95-2952m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.146039 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prkxz\" (UniqueName: \"kubernetes.io/projected/26b61717-1239-4782-bc75-6b0eece01c14-kube-api-access-prkxz\") pod \"dnsmasq-dns-85ff748b95-2952m\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " pod="openstack/dnsmasq-dns-85ff748b95-2952m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.274613 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f8a567c-2577-4c66-a75e-e43f7cd66000-config-data\") pod \"barbican-api-5f9b44b4db-xxc6m\" (UID: \"9f8a567c-2577-4c66-a75e-e43f7cd66000\") " pod="openstack/barbican-api-5f9b44b4db-xxc6m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.275203 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-2952m\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " pod="openstack/dnsmasq-dns-85ff748b95-2952m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.275247 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-2952m\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " pod="openstack/dnsmasq-dns-85ff748b95-2952m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.275294 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f8a567c-2577-4c66-a75e-e43f7cd66000-config-data-custom\") pod \"barbican-api-5f9b44b4db-xxc6m\" (UID: \"9f8a567c-2577-4c66-a75e-e43f7cd66000\") " pod="openstack/barbican-api-5f9b44b4db-xxc6m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.275324 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-2952m\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " pod="openstack/dnsmasq-dns-85ff748b95-2952m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.275485 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f8a567c-2577-4c66-a75e-e43f7cd66000-logs\") pod \"barbican-api-5f9b44b4db-xxc6m\" (UID: \"9f8a567c-2577-4c66-a75e-e43f7cd66000\") " pod="openstack/barbican-api-5f9b44b4db-xxc6m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.275526 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-config\") pod \"dnsmasq-dns-85ff748b95-2952m\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " pod="openstack/dnsmasq-dns-85ff748b95-2952m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.275663 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f8a567c-2577-4c66-a75e-e43f7cd66000-combined-ca-bundle\") pod \"barbican-api-5f9b44b4db-xxc6m\" (UID: \"9f8a567c-2577-4c66-a75e-e43f7cd66000\") " pod="openstack/barbican-api-5f9b44b4db-xxc6m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.275717 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp4c2\" (UniqueName: \"kubernetes.io/projected/9f8a567c-2577-4c66-a75e-e43f7cd66000-kube-api-access-jp4c2\") pod \"barbican-api-5f9b44b4db-xxc6m\" (UID: \"9f8a567c-2577-4c66-a75e-e43f7cd66000\") " pod="openstack/barbican-api-5f9b44b4db-xxc6m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.275825 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-dns-svc\") pod \"dnsmasq-dns-85ff748b95-2952m\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " pod="openstack/dnsmasq-dns-85ff748b95-2952m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.275919 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prkxz\" (UniqueName: \"kubernetes.io/projected/26b61717-1239-4782-bc75-6b0eece01c14-kube-api-access-prkxz\") pod \"dnsmasq-dns-85ff748b95-2952m\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " pod="openstack/dnsmasq-dns-85ff748b95-2952m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.281525 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-2952m\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " pod="openstack/dnsmasq-dns-85ff748b95-2952m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.299193 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-config\") pod \"dnsmasq-dns-85ff748b95-2952m\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " pod="openstack/dnsmasq-dns-85ff748b95-2952m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.299584 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f8a567c-2577-4c66-a75e-e43f7cd66000-logs\") pod \"barbican-api-5f9b44b4db-xxc6m\" (UID: \"9f8a567c-2577-4c66-a75e-e43f7cd66000\") " pod="openstack/barbican-api-5f9b44b4db-xxc6m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.300280 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-2952m\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " pod="openstack/dnsmasq-dns-85ff748b95-2952m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.313733 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f8a567c-2577-4c66-a75e-e43f7cd66000-config-data\") pod \"barbican-api-5f9b44b4db-xxc6m\" (UID: \"9f8a567c-2577-4c66-a75e-e43f7cd66000\") " pod="openstack/barbican-api-5f9b44b4db-xxc6m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.313867 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-dns-svc\") pod \"dnsmasq-dns-85ff748b95-2952m\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " pod="openstack/dnsmasq-dns-85ff748b95-2952m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.328587 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-2952m\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " pod="openstack/dnsmasq-dns-85ff748b95-2952m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.334649 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f8a567c-2577-4c66-a75e-e43f7cd66000-combined-ca-bundle\") pod \"barbican-api-5f9b44b4db-xxc6m\" (UID: \"9f8a567c-2577-4c66-a75e-e43f7cd66000\") " pod="openstack/barbican-api-5f9b44b4db-xxc6m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.354487 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp4c2\" (UniqueName: \"kubernetes.io/projected/9f8a567c-2577-4c66-a75e-e43f7cd66000-kube-api-access-jp4c2\") pod \"barbican-api-5f9b44b4db-xxc6m\" (UID: \"9f8a567c-2577-4c66-a75e-e43f7cd66000\") " pod="openstack/barbican-api-5f9b44b4db-xxc6m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.355180 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f8a567c-2577-4c66-a75e-e43f7cd66000-config-data-custom\") pod \"barbican-api-5f9b44b4db-xxc6m\" (UID: \"9f8a567c-2577-4c66-a75e-e43f7cd66000\") " pod="openstack/barbican-api-5f9b44b4db-xxc6m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.381909 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prkxz\" (UniqueName: \"kubernetes.io/projected/26b61717-1239-4782-bc75-6b0eece01c14-kube-api-access-prkxz\") pod \"dnsmasq-dns-85ff748b95-2952m\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " pod="openstack/dnsmasq-dns-85ff748b95-2952m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.574539 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-2952m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.621409 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f9b44b4db-xxc6m" Dec 09 10:12:09 crc kubenswrapper[4824]: I1209 10:12:09.849034 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-55b96fff66-brzz9"] Dec 09 10:12:10 crc kubenswrapper[4824]: E1209 10:12:10.034934 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad8cfdb4_f7e5_4b30_a1b8_a680b111f788.slice/crio-conmon-c0650433deec5abea39f2881e472191bd159aadb67a9c737ba5bd6be7c0a63c9.scope\": RecentStats: unable to find data in memory cache]" Dec 09 10:12:10 crc kubenswrapper[4824]: I1209 10:12:10.053234 4824 generic.go:334] "Generic (PLEG): container finished" podID="ad8cfdb4-f7e5-4b30-a1b8-a680b111f788" containerID="c0650433deec5abea39f2881e472191bd159aadb67a9c737ba5bd6be7c0a63c9" exitCode=0 Dec 09 10:12:10 crc kubenswrapper[4824]: I1209 10:12:10.053403 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 10:12:10 crc kubenswrapper[4824]: I1209 10:12:10.053418 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 10:12:10 crc kubenswrapper[4824]: I1209 10:12:10.058753 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff1309c3-9703-4b9c-8ec7-e9fdeee87b31" path="/var/lib/kubelet/pods/ff1309c3-9703-4b9c-8ec7-e9fdeee87b31/volumes" Dec 09 10:12:10 crc kubenswrapper[4824]: I1209 10:12:10.061340 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-55b96fff66-brzz9" event={"ID":"d72ff513-d8d9-4191-92d4-7824fca18f32","Type":"ContainerStarted","Data":"70dc6d2e95195b3e192fccef3cbc251bb2c23e2d79d35fc530e170a62dcc5457"} Dec 09 10:12:10 crc kubenswrapper[4824]: I1209 10:12:10.061368 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lhlfr" event={"ID":"ad8cfdb4-f7e5-4b30-a1b8-a680b111f788","Type":"ContainerDied","Data":"c0650433deec5abea39f2881e472191bd159aadb67a9c737ba5bd6be7c0a63c9"} Dec 09 10:12:10 crc kubenswrapper[4824]: I1209 10:12:10.424062 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-9c7879457-2fmzw"] Dec 09 10:12:10 crc kubenswrapper[4824]: I1209 10:12:10.438498 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-77bcf58d55-bshs5"] Dec 09 10:12:10 crc kubenswrapper[4824]: I1209 10:12:10.601838 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7c7958c67b-wr8zx"] Dec 09 10:12:10 crc kubenswrapper[4824]: W1209 10:12:10.631877 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38c32e83_50e7_4f8b_8bbf_26adebff1b82.slice/crio-25ba18531c4abd1cdc17f147a2d38fdd5fe508b5454a6d7b7c96f30c33618c00 WatchSource:0}: Error finding container 25ba18531c4abd1cdc17f147a2d38fdd5fe508b5454a6d7b7c96f30c33618c00: Status 404 returned error can't find the container with id 25ba18531c4abd1cdc17f147a2d38fdd5fe508b5454a6d7b7c96f30c33618c00 Dec 09 10:12:10 crc kubenswrapper[4824]: I1209 10:12:10.748531 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-2952m"] Dec 09 10:12:10 crc kubenswrapper[4824]: I1209 10:12:10.770946 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5f9b44b4db-xxc6m"] Dec 09 10:12:10 crc kubenswrapper[4824]: W1209 10:12:10.783960 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26b61717_1239_4782_bc75_6b0eece01c14.slice/crio-c52022598eb16d1fb834933ce2c44550354a25075891170be311b4944b156302 WatchSource:0}: Error finding container c52022598eb16d1fb834933ce2c44550354a25075891170be311b4944b156302: Status 404 returned error can't find the container with id c52022598eb16d1fb834933ce2c44550354a25075891170be311b4944b156302 Dec 09 10:12:11 crc kubenswrapper[4824]: I1209 10:12:11.089933 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-77bcf58d55-bshs5" event={"ID":"555ef974-afb9-4455-8a15-c728200be8a2","Type":"ContainerStarted","Data":"4315423fd6e5e2d4049eb1a1321a75351f87295c0c039a2ae45db12e49b581f9"} Dec 09 10:12:11 crc kubenswrapper[4824]: I1209 10:12:11.108407 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lhlfr" event={"ID":"ad8cfdb4-f7e5-4b30-a1b8-a680b111f788","Type":"ContainerStarted","Data":"d2821f5dfa01bb037746d395dcb971a50b3480320fac295f2df8ad81c9fe1e90"} Dec 09 10:12:11 crc kubenswrapper[4824]: I1209 10:12:11.117738 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9c7879457-2fmzw" event={"ID":"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7","Type":"ContainerStarted","Data":"01779be8c7291db7676c77985367964b6a543b5ed2ba0af30547363a92bf4797"} Dec 09 10:12:11 crc kubenswrapper[4824]: I1209 10:12:11.117827 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9c7879457-2fmzw" event={"ID":"9dcb334f-65bf-4f84-88a0-24ca56b4a8c7","Type":"ContainerStarted","Data":"08d85e6652bff039b5174f6d0082ccc36b87f480c626fdb64e822a7b77b8e319"} Dec 09 10:12:11 crc kubenswrapper[4824]: I1209 10:12:11.117874 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:11 crc kubenswrapper[4824]: I1209 10:12:11.119394 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f9b44b4db-xxc6m" event={"ID":"9f8a567c-2577-4c66-a75e-e43f7cd66000","Type":"ContainerStarted","Data":"04e9bcf61ce76c0a2526dca38ba02d2faf65851ef66dbf13fd8b4a6a18f6acb8"} Dec 09 10:12:11 crc kubenswrapper[4824]: I1209 10:12:11.123140 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-2952m" event={"ID":"26b61717-1239-4782-bc75-6b0eece01c14","Type":"ContainerStarted","Data":"c52022598eb16d1fb834933ce2c44550354a25075891170be311b4944b156302"} Dec 09 10:12:11 crc kubenswrapper[4824]: I1209 10:12:11.142742 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7c7958c67b-wr8zx" event={"ID":"38c32e83-50e7-4f8b-8bbf-26adebff1b82","Type":"ContainerStarted","Data":"25ba18531c4abd1cdc17f147a2d38fdd5fe508b5454a6d7b7c96f30c33618c00"} Dec 09 10:12:11 crc kubenswrapper[4824]: I1209 10:12:11.171180 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-9c7879457-2fmzw" podStartSLOduration=3.171162941 podStartE2EDuration="3.171162941s" podCreationTimestamp="2025-12-09 10:12:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:12:11.168244959 +0000 UTC m=+1487.502749626" watchObservedRunningTime="2025-12-09 10:12:11.171162941 +0000 UTC m=+1487.505667608" Dec 09 10:12:11 crc kubenswrapper[4824]: I1209 10:12:11.181403 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lhlfr" podStartSLOduration=5.281671843 podStartE2EDuration="14.181385756s" podCreationTimestamp="2025-12-09 10:11:57 +0000 UTC" firstStartedPulling="2025-12-09 10:12:01.781651799 +0000 UTC m=+1478.116156476" lastFinishedPulling="2025-12-09 10:12:10.681365722 +0000 UTC m=+1487.015870389" observedRunningTime="2025-12-09 10:12:11.142909732 +0000 UTC m=+1487.477414419" watchObservedRunningTime="2025-12-09 10:12:11.181385756 +0000 UTC m=+1487.515890423" Dec 09 10:12:12 crc kubenswrapper[4824]: I1209 10:12:12.193136 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f9b44b4db-xxc6m" event={"ID":"9f8a567c-2577-4c66-a75e-e43f7cd66000","Type":"ContainerStarted","Data":"5ea3d7e5055f26b89248f5118b0972582d7d96bea9dedca1d4da15bbb311e5d7"} Dec 09 10:12:12 crc kubenswrapper[4824]: I1209 10:12:12.193914 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f9b44b4db-xxc6m" event={"ID":"9f8a567c-2577-4c66-a75e-e43f7cd66000","Type":"ContainerStarted","Data":"b5cf173a26f1977cf61d3b4184fa796463dc2a63ac661f657cca3f3e873af83f"} Dec 09 10:12:12 crc kubenswrapper[4824]: I1209 10:12:12.195034 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5f9b44b4db-xxc6m" Dec 09 10:12:12 crc kubenswrapper[4824]: I1209 10:12:12.195090 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5f9b44b4db-xxc6m" Dec 09 10:12:12 crc kubenswrapper[4824]: I1209 10:12:12.215549 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b61717-1239-4782-bc75-6b0eece01c14" containerID="ebcec85021517dc9f266f1c774c51ff8974b3d9223f06e0739ebf4773c604928" exitCode=0 Dec 09 10:12:12 crc kubenswrapper[4824]: I1209 10:12:12.215664 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-2952m" event={"ID":"26b61717-1239-4782-bc75-6b0eece01c14","Type":"ContainerDied","Data":"ebcec85021517dc9f266f1c774c51ff8974b3d9223f06e0739ebf4773c604928"} Dec 09 10:12:12 crc kubenswrapper[4824]: I1209 10:12:12.218619 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5f9b44b4db-xxc6m" podStartSLOduration=4.218602501 podStartE2EDuration="4.218602501s" podCreationTimestamp="2025-12-09 10:12:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:12:12.21825222 +0000 UTC m=+1488.552756897" watchObservedRunningTime="2025-12-09 10:12:12.218602501 +0000 UTC m=+1488.553107168" Dec 09 10:12:12 crc kubenswrapper[4824]: I1209 10:12:12.240759 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7c7958c67b-wr8zx" event={"ID":"38c32e83-50e7-4f8b-8bbf-26adebff1b82","Type":"ContainerStarted","Data":"920c7d6a7569241d5825ec023ed8f7f5db47de15efedee9d8e73c86d7b5066fb"} Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.088227 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-867bb6d64d-9slcv"] Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.091272 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.095243 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.095346 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.126499 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-867bb6d64d-9slcv"] Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.143628 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4dc240ff-1bf1-464e-8733-59d0ae2fc4d2-internal-tls-certs\") pod \"barbican-api-867bb6d64d-9slcv\" (UID: \"4dc240ff-1bf1-464e-8733-59d0ae2fc4d2\") " pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.143704 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4dc240ff-1bf1-464e-8733-59d0ae2fc4d2-public-tls-certs\") pod \"barbican-api-867bb6d64d-9slcv\" (UID: \"4dc240ff-1bf1-464e-8733-59d0ae2fc4d2\") " pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.143735 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4dc240ff-1bf1-464e-8733-59d0ae2fc4d2-logs\") pod \"barbican-api-867bb6d64d-9slcv\" (UID: \"4dc240ff-1bf1-464e-8733-59d0ae2fc4d2\") " pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.143866 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4dc240ff-1bf1-464e-8733-59d0ae2fc4d2-config-data-custom\") pod \"barbican-api-867bb6d64d-9slcv\" (UID: \"4dc240ff-1bf1-464e-8733-59d0ae2fc4d2\") " pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.143924 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dc240ff-1bf1-464e-8733-59d0ae2fc4d2-combined-ca-bundle\") pod \"barbican-api-867bb6d64d-9slcv\" (UID: \"4dc240ff-1bf1-464e-8733-59d0ae2fc4d2\") " pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.144068 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9cb7\" (UniqueName: \"kubernetes.io/projected/4dc240ff-1bf1-464e-8733-59d0ae2fc4d2-kube-api-access-l9cb7\") pod \"barbican-api-867bb6d64d-9slcv\" (UID: \"4dc240ff-1bf1-464e-8733-59d0ae2fc4d2\") " pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.144092 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4dc240ff-1bf1-464e-8733-59d0ae2fc4d2-config-data\") pod \"barbican-api-867bb6d64d-9slcv\" (UID: \"4dc240ff-1bf1-464e-8733-59d0ae2fc4d2\") " pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.249144 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4dc240ff-1bf1-464e-8733-59d0ae2fc4d2-public-tls-certs\") pod \"barbican-api-867bb6d64d-9slcv\" (UID: \"4dc240ff-1bf1-464e-8733-59d0ae2fc4d2\") " pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.249222 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4dc240ff-1bf1-464e-8733-59d0ae2fc4d2-logs\") pod \"barbican-api-867bb6d64d-9slcv\" (UID: \"4dc240ff-1bf1-464e-8733-59d0ae2fc4d2\") " pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.249303 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4dc240ff-1bf1-464e-8733-59d0ae2fc4d2-config-data-custom\") pod \"barbican-api-867bb6d64d-9slcv\" (UID: \"4dc240ff-1bf1-464e-8733-59d0ae2fc4d2\") " pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.249358 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dc240ff-1bf1-464e-8733-59d0ae2fc4d2-combined-ca-bundle\") pod \"barbican-api-867bb6d64d-9slcv\" (UID: \"4dc240ff-1bf1-464e-8733-59d0ae2fc4d2\") " pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.249472 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9cb7\" (UniqueName: \"kubernetes.io/projected/4dc240ff-1bf1-464e-8733-59d0ae2fc4d2-kube-api-access-l9cb7\") pod \"barbican-api-867bb6d64d-9slcv\" (UID: \"4dc240ff-1bf1-464e-8733-59d0ae2fc4d2\") " pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.249496 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4dc240ff-1bf1-464e-8733-59d0ae2fc4d2-config-data\") pod \"barbican-api-867bb6d64d-9slcv\" (UID: \"4dc240ff-1bf1-464e-8733-59d0ae2fc4d2\") " pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.249593 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4dc240ff-1bf1-464e-8733-59d0ae2fc4d2-internal-tls-certs\") pod \"barbican-api-867bb6d64d-9slcv\" (UID: \"4dc240ff-1bf1-464e-8733-59d0ae2fc4d2\") " pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.256246 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4dc240ff-1bf1-464e-8733-59d0ae2fc4d2-logs\") pod \"barbican-api-867bb6d64d-9slcv\" (UID: \"4dc240ff-1bf1-464e-8733-59d0ae2fc4d2\") " pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.257588 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dc240ff-1bf1-464e-8733-59d0ae2fc4d2-combined-ca-bundle\") pod \"barbican-api-867bb6d64d-9slcv\" (UID: \"4dc240ff-1bf1-464e-8733-59d0ae2fc4d2\") " pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.258801 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4dc240ff-1bf1-464e-8733-59d0ae2fc4d2-config-data-custom\") pod \"barbican-api-867bb6d64d-9slcv\" (UID: \"4dc240ff-1bf1-464e-8733-59d0ae2fc4d2\") " pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.259615 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4dc240ff-1bf1-464e-8733-59d0ae2fc4d2-internal-tls-certs\") pod \"barbican-api-867bb6d64d-9slcv\" (UID: \"4dc240ff-1bf1-464e-8733-59d0ae2fc4d2\") " pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.266326 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4dc240ff-1bf1-464e-8733-59d0ae2fc4d2-public-tls-certs\") pod \"barbican-api-867bb6d64d-9slcv\" (UID: \"4dc240ff-1bf1-464e-8733-59d0ae2fc4d2\") " pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.270021 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4dc240ff-1bf1-464e-8733-59d0ae2fc4d2-config-data\") pod \"barbican-api-867bb6d64d-9slcv\" (UID: \"4dc240ff-1bf1-464e-8733-59d0ae2fc4d2\") " pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.279544 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9cb7\" (UniqueName: \"kubernetes.io/projected/4dc240ff-1bf1-464e-8733-59d0ae2fc4d2-kube-api-access-l9cb7\") pod \"barbican-api-867bb6d64d-9slcv\" (UID: \"4dc240ff-1bf1-464e-8733-59d0ae2fc4d2\") " pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.301365 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-2952m" event={"ID":"26b61717-1239-4782-bc75-6b0eece01c14","Type":"ContainerStarted","Data":"bc738525e61a76ad921d11262bd0345eb8616de11b3587e7a1de703692188cda"} Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.301517 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85ff748b95-2952m" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.311678 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7c7958c67b-wr8zx" event={"ID":"38c32e83-50e7-4f8b-8bbf-26adebff1b82","Type":"ContainerStarted","Data":"9479903881dca38ce49333f11f4bf4d95027a52664db3a7f29f135a4fa7bd1ee"} Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.311724 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.311753 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.421645 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.443465 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85ff748b95-2952m" podStartSLOduration=5.443435757 podStartE2EDuration="5.443435757s" podCreationTimestamp="2025-12-09 10:12:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:12:13.43033561 +0000 UTC m=+1489.764840277" watchObservedRunningTime="2025-12-09 10:12:13.443435757 +0000 UTC m=+1489.777940414" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.486077 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7c7958c67b-wr8zx" podStartSLOduration=5.486045833 podStartE2EDuration="5.486045833s" podCreationTimestamp="2025-12-09 10:12:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:12:13.460606623 +0000 UTC m=+1489.795111280" watchObservedRunningTime="2025-12-09 10:12:13.486045833 +0000 UTC m=+1489.820550510" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.699680 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.700184 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.704962 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.707794 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.707929 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 10:12:13 crc kubenswrapper[4824]: I1209 10:12:13.710395 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 09 10:12:16 crc kubenswrapper[4824]: I1209 10:12:16.836358 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-867bb6d64d-9slcv"] Dec 09 10:12:17 crc kubenswrapper[4824]: I1209 10:12:17.423249 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-55b96fff66-brzz9" event={"ID":"d72ff513-d8d9-4191-92d4-7824fca18f32","Type":"ContainerStarted","Data":"006b98f55544faeb9995affc03e1dfb2d767c744a324a007d5124aea19d93f19"} Dec 09 10:12:17 crc kubenswrapper[4824]: I1209 10:12:17.423684 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-55b96fff66-brzz9" event={"ID":"d72ff513-d8d9-4191-92d4-7824fca18f32","Type":"ContainerStarted","Data":"1cc1d4554da1285e17345181e9e921151bff7970c17190bfe94f852dac68ad07"} Dec 09 10:12:17 crc kubenswrapper[4824]: I1209 10:12:17.426248 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-77bcf58d55-bshs5" event={"ID":"555ef974-afb9-4455-8a15-c728200be8a2","Type":"ContainerStarted","Data":"666ab1e035a8417a26f2ecd83a6d89f43ab8e791a490317f8edec5d0a86fd826"} Dec 09 10:12:17 crc kubenswrapper[4824]: I1209 10:12:17.426308 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-77bcf58d55-bshs5" event={"ID":"555ef974-afb9-4455-8a15-c728200be8a2","Type":"ContainerStarted","Data":"086ae2ebeb5ab637bb6408a85dc87b59cf64e9619651790f9adadecfe95a47ef"} Dec 09 10:12:17 crc kubenswrapper[4824]: I1209 10:12:17.429399 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-867bb6d64d-9slcv" event={"ID":"4dc240ff-1bf1-464e-8733-59d0ae2fc4d2","Type":"ContainerStarted","Data":"fc5f802bac1c1cac14a92072b11403ab0529beada9fd3a7ae59ed0e9bc5420b0"} Dec 09 10:12:17 crc kubenswrapper[4824]: I1209 10:12:17.429457 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-867bb6d64d-9slcv" event={"ID":"4dc240ff-1bf1-464e-8733-59d0ae2fc4d2","Type":"ContainerStarted","Data":"9be766abe5527cfa4b1c6a9a7fe756bac634dcbebb78eae7985270791c0941af"} Dec 09 10:12:17 crc kubenswrapper[4824]: I1209 10:12:17.444772 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-55b96fff66-brzz9" podStartSLOduration=3.194713831 podStartE2EDuration="9.444744376s" podCreationTimestamp="2025-12-09 10:12:08 +0000 UTC" firstStartedPulling="2025-12-09 10:12:09.910735293 +0000 UTC m=+1486.245239960" lastFinishedPulling="2025-12-09 10:12:16.160765838 +0000 UTC m=+1492.495270505" observedRunningTime="2025-12-09 10:12:17.441820563 +0000 UTC m=+1493.776325250" watchObservedRunningTime="2025-12-09 10:12:17.444744376 +0000 UTC m=+1493.779249043" Dec 09 10:12:17 crc kubenswrapper[4824]: I1209 10:12:17.480825 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-77bcf58d55-bshs5" podStartSLOduration=3.782807989 podStartE2EDuration="9.480803023s" podCreationTimestamp="2025-12-09 10:12:08 +0000 UTC" firstStartedPulling="2025-12-09 10:12:10.462739423 +0000 UTC m=+1486.797244090" lastFinishedPulling="2025-12-09 10:12:16.160734467 +0000 UTC m=+1492.495239124" observedRunningTime="2025-12-09 10:12:17.474639257 +0000 UTC m=+1493.809143944" watchObservedRunningTime="2025-12-09 10:12:17.480803023 +0000 UTC m=+1493.815307690" Dec 09 10:12:18 crc kubenswrapper[4824]: I1209 10:12:18.234475 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lhlfr" Dec 09 10:12:18 crc kubenswrapper[4824]: I1209 10:12:18.236140 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lhlfr" Dec 09 10:12:18 crc kubenswrapper[4824]: I1209 10:12:18.463659 4824 generic.go:334] "Generic (PLEG): container finished" podID="405d0a5e-fd62-4146-bfb9-96ff6c077836" containerID="40ff06ab3f0f7bbf01219bb470c28376ccf91ab46f772e08dc41ca02ca67a752" exitCode=0 Dec 09 10:12:18 crc kubenswrapper[4824]: I1209 10:12:18.464932 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-vfnsw" event={"ID":"405d0a5e-fd62-4146-bfb9-96ff6c077836","Type":"ContainerDied","Data":"40ff06ab3f0f7bbf01219bb470c28376ccf91ab46f772e08dc41ca02ca67a752"} Dec 09 10:12:19 crc kubenswrapper[4824]: I1209 10:12:19.499422 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lhlfr" podUID="ad8cfdb4-f7e5-4b30-a1b8-a680b111f788" containerName="registry-server" probeResult="failure" output=< Dec 09 10:12:19 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 10:12:19 crc kubenswrapper[4824]: > Dec 09 10:12:19 crc kubenswrapper[4824]: I1209 10:12:19.583093 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85ff748b95-2952m" Dec 09 10:12:19 crc kubenswrapper[4824]: I1209 10:12:19.719481 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-pmxwc"] Dec 09 10:12:19 crc kubenswrapper[4824]: I1209 10:12:19.720128 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" podUID="ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780" containerName="dnsmasq-dns" containerID="cri-o://20a51d23f331fb43fd5bb76eccbc3e81fdb36fcc151c9f8fb0f57733abeca4bf" gracePeriod=10 Dec 09 10:12:20 crc kubenswrapper[4824]: I1209 10:12:20.185931 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" podUID="ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.192:5353: connect: connection refused" Dec 09 10:12:20 crc kubenswrapper[4824]: I1209 10:12:20.340260 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5855dddcbd-6qlwr" Dec 09 10:12:21 crc kubenswrapper[4824]: I1209 10:12:21.011220 4824 generic.go:334] "Generic (PLEG): container finished" podID="ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780" containerID="20a51d23f331fb43fd5bb76eccbc3e81fdb36fcc151c9f8fb0f57733abeca4bf" exitCode=0 Dec 09 10:12:21 crc kubenswrapper[4824]: I1209 10:12:21.011680 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" event={"ID":"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780","Type":"ContainerDied","Data":"20a51d23f331fb43fd5bb76eccbc3e81fdb36fcc151c9f8fb0f57733abeca4bf"} Dec 09 10:12:21 crc kubenswrapper[4824]: I1209 10:12:21.016066 4824 generic.go:334] "Generic (PLEG): container finished" podID="dbceafb7-272a-45fc-9b56-df631986f90a" containerID="751410b43bd034373b18f9792b5d576ad6b6aa7d073a4a19569c94413f9c45c0" exitCode=0 Dec 09 10:12:21 crc kubenswrapper[4824]: I1209 10:12:21.016110 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-jghjq" event={"ID":"dbceafb7-272a-45fc-9b56-df631986f90a","Type":"ContainerDied","Data":"751410b43bd034373b18f9792b5d576ad6b6aa7d073a4a19569c94413f9c45c0"} Dec 09 10:12:21 crc kubenswrapper[4824]: I1209 10:12:21.577344 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5f9b44b4db-xxc6m" podUID="9f8a567c-2577-4c66-a75e-e43f7cd66000" containerName="barbican-api" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 10:12:22 crc kubenswrapper[4824]: I1209 10:12:22.110674 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5f9b44b4db-xxc6m" Dec 09 10:12:22 crc kubenswrapper[4824]: I1209 10:12:22.232426 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5f9b44b4db-xxc6m" Dec 09 10:12:22 crc kubenswrapper[4824]: I1209 10:12:22.991973 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7d56cf488f-6wtnd" Dec 09 10:12:23 crc kubenswrapper[4824]: I1209 10:12:23.089500 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5855dddcbd-6qlwr"] Dec 09 10:12:23 crc kubenswrapper[4824]: I1209 10:12:23.089793 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5855dddcbd-6qlwr" podUID="95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3" containerName="neutron-api" containerID="cri-o://39ac59a1f6fbf896b4c2d08f9997e40dec30a053bbbcc01bf095c383bce4b7af" gracePeriod=30 Dec 09 10:12:23 crc kubenswrapper[4824]: I1209 10:12:23.089943 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5855dddcbd-6qlwr" podUID="95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3" containerName="neutron-httpd" containerID="cri-o://8e6ba1b57bd352f45b14b90495296a5ad4334e5ba289f62bcfb6b62fdd0b9de2" gracePeriod=30 Dec 09 10:12:24 crc kubenswrapper[4824]: I1209 10:12:24.034243 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-vfnsw" Dec 09 10:12:24 crc kubenswrapper[4824]: I1209 10:12:24.061355 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/405d0a5e-fd62-4146-bfb9-96ff6c077836-combined-ca-bundle\") pod \"405d0a5e-fd62-4146-bfb9-96ff6c077836\" (UID: \"405d0a5e-fd62-4146-bfb9-96ff6c077836\") " Dec 09 10:12:24 crc kubenswrapper[4824]: I1209 10:12:24.061440 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2csn\" (UniqueName: \"kubernetes.io/projected/405d0a5e-fd62-4146-bfb9-96ff6c077836-kube-api-access-p2csn\") pod \"405d0a5e-fd62-4146-bfb9-96ff6c077836\" (UID: \"405d0a5e-fd62-4146-bfb9-96ff6c077836\") " Dec 09 10:12:24 crc kubenswrapper[4824]: I1209 10:12:24.061475 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/405d0a5e-fd62-4146-bfb9-96ff6c077836-config-data\") pod \"405d0a5e-fd62-4146-bfb9-96ff6c077836\" (UID: \"405d0a5e-fd62-4146-bfb9-96ff6c077836\") " Dec 09 10:12:24 crc kubenswrapper[4824]: I1209 10:12:24.075589 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-vfnsw" event={"ID":"405d0a5e-fd62-4146-bfb9-96ff6c077836","Type":"ContainerDied","Data":"e09a391a1ba56ef305ee8e1a594b4b370dc65c7068801fc2b2a7eba8a2015ce8"} Dec 09 10:12:24 crc kubenswrapper[4824]: I1209 10:12:24.075635 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e09a391a1ba56ef305ee8e1a594b4b370dc65c7068801fc2b2a7eba8a2015ce8" Dec 09 10:12:24 crc kubenswrapper[4824]: I1209 10:12:24.075719 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-vfnsw" Dec 09 10:12:24 crc kubenswrapper[4824]: I1209 10:12:24.078457 4824 generic.go:334] "Generic (PLEG): container finished" podID="95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3" containerID="8e6ba1b57bd352f45b14b90495296a5ad4334e5ba289f62bcfb6b62fdd0b9de2" exitCode=0 Dec 09 10:12:24 crc kubenswrapper[4824]: I1209 10:12:24.078593 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5855dddcbd-6qlwr" event={"ID":"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3","Type":"ContainerDied","Data":"8e6ba1b57bd352f45b14b90495296a5ad4334e5ba289f62bcfb6b62fdd0b9de2"} Dec 09 10:12:24 crc kubenswrapper[4824]: I1209 10:12:24.085870 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/405d0a5e-fd62-4146-bfb9-96ff6c077836-kube-api-access-p2csn" (OuterVolumeSpecName: "kube-api-access-p2csn") pod "405d0a5e-fd62-4146-bfb9-96ff6c077836" (UID: "405d0a5e-fd62-4146-bfb9-96ff6c077836"). InnerVolumeSpecName "kube-api-access-p2csn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:12:24 crc kubenswrapper[4824]: I1209 10:12:24.117606 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/405d0a5e-fd62-4146-bfb9-96ff6c077836-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "405d0a5e-fd62-4146-bfb9-96ff6c077836" (UID: "405d0a5e-fd62-4146-bfb9-96ff6c077836"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:24 crc kubenswrapper[4824]: I1209 10:12:24.163765 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2csn\" (UniqueName: \"kubernetes.io/projected/405d0a5e-fd62-4146-bfb9-96ff6c077836-kube-api-access-p2csn\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:24 crc kubenswrapper[4824]: I1209 10:12:24.163820 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/405d0a5e-fd62-4146-bfb9-96ff6c077836-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:24 crc kubenswrapper[4824]: I1209 10:12:24.199806 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/405d0a5e-fd62-4146-bfb9-96ff6c077836-config-data" (OuterVolumeSpecName: "config-data") pod "405d0a5e-fd62-4146-bfb9-96ff6c077836" (UID: "405d0a5e-fd62-4146-bfb9-96ff6c077836"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:24 crc kubenswrapper[4824]: I1209 10:12:24.265425 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/405d0a5e-fd62-4146-bfb9-96ff6c077836-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:25 crc kubenswrapper[4824]: I1209 10:12:25.184062 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" podUID="ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.192:5353: connect: connection refused" Dec 09 10:12:26 crc kubenswrapper[4824]: I1209 10:12:26.875242 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-jghjq" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.037436 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbceafb7-272a-45fc-9b56-df631986f90a-config-data\") pod \"dbceafb7-272a-45fc-9b56-df631986f90a\" (UID: \"dbceafb7-272a-45fc-9b56-df631986f90a\") " Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.038084 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dbceafb7-272a-45fc-9b56-df631986f90a-db-sync-config-data\") pod \"dbceafb7-272a-45fc-9b56-df631986f90a\" (UID: \"dbceafb7-272a-45fc-9b56-df631986f90a\") " Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.038160 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbceafb7-272a-45fc-9b56-df631986f90a-scripts\") pod \"dbceafb7-272a-45fc-9b56-df631986f90a\" (UID: \"dbceafb7-272a-45fc-9b56-df631986f90a\") " Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.038321 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dbceafb7-272a-45fc-9b56-df631986f90a-etc-machine-id\") pod \"dbceafb7-272a-45fc-9b56-df631986f90a\" (UID: \"dbceafb7-272a-45fc-9b56-df631986f90a\") " Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.038405 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dbceafb7-272a-45fc-9b56-df631986f90a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "dbceafb7-272a-45fc-9b56-df631986f90a" (UID: "dbceafb7-272a-45fc-9b56-df631986f90a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.043143 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwnjz\" (UniqueName: \"kubernetes.io/projected/dbceafb7-272a-45fc-9b56-df631986f90a-kube-api-access-lwnjz\") pod \"dbceafb7-272a-45fc-9b56-df631986f90a\" (UID: \"dbceafb7-272a-45fc-9b56-df631986f90a\") " Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.043971 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbceafb7-272a-45fc-9b56-df631986f90a-combined-ca-bundle\") pod \"dbceafb7-272a-45fc-9b56-df631986f90a\" (UID: \"dbceafb7-272a-45fc-9b56-df631986f90a\") " Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.045829 4824 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dbceafb7-272a-45fc-9b56-df631986f90a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.046560 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbceafb7-272a-45fc-9b56-df631986f90a-scripts" (OuterVolumeSpecName: "scripts") pod "dbceafb7-272a-45fc-9b56-df631986f90a" (UID: "dbceafb7-272a-45fc-9b56-df631986f90a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.051374 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbceafb7-272a-45fc-9b56-df631986f90a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "dbceafb7-272a-45fc-9b56-df631986f90a" (UID: "dbceafb7-272a-45fc-9b56-df631986f90a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.066483 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbceafb7-272a-45fc-9b56-df631986f90a-kube-api-access-lwnjz" (OuterVolumeSpecName: "kube-api-access-lwnjz") pod "dbceafb7-272a-45fc-9b56-df631986f90a" (UID: "dbceafb7-272a-45fc-9b56-df631986f90a"). InnerVolumeSpecName "kube-api-access-lwnjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.089526 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbceafb7-272a-45fc-9b56-df631986f90a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dbceafb7-272a-45fc-9b56-df631986f90a" (UID: "dbceafb7-272a-45fc-9b56-df631986f90a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.143282 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbceafb7-272a-45fc-9b56-df631986f90a-config-data" (OuterVolumeSpecName: "config-data") pod "dbceafb7-272a-45fc-9b56-df631986f90a" (UID: "dbceafb7-272a-45fc-9b56-df631986f90a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.143712 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-jghjq" event={"ID":"dbceafb7-272a-45fc-9b56-df631986f90a","Type":"ContainerDied","Data":"1c49fa2af8410be9172d0e097fdd5701dbea33fbc41b5f7c18ae4593604a9f73"} Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.143747 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c49fa2af8410be9172d0e097fdd5701dbea33fbc41b5f7c18ae4593604a9f73" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.143853 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-jghjq" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.149669 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbceafb7-272a-45fc-9b56-df631986f90a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.149727 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbceafb7-272a-45fc-9b56-df631986f90a-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.149741 4824 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dbceafb7-272a-45fc-9b56-df631986f90a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.149753 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbceafb7-272a-45fc-9b56-df631986f90a-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.149813 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwnjz\" (UniqueName: \"kubernetes.io/projected/dbceafb7-272a-45fc-9b56-df631986f90a-kube-api-access-lwnjz\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:27 crc kubenswrapper[4824]: E1209 10:12:27.159577 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="232c4238-3f59-4e05-ac75-8746ab0a0069" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.174846 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.251018 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-ovsdbserver-sb\") pod \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\" (UID: \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\") " Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.251217 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-ovsdbserver-nb\") pod \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\" (UID: \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\") " Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.251275 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-dns-swift-storage-0\") pod \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\" (UID: \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\") " Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.251386 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-config\") pod \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\" (UID: \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\") " Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.251414 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-dns-svc\") pod \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\" (UID: \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\") " Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.251506 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzwkk\" (UniqueName: \"kubernetes.io/projected/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-kube-api-access-tzwkk\") pod \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\" (UID: \"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780\") " Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.274346 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-kube-api-access-tzwkk" (OuterVolumeSpecName: "kube-api-access-tzwkk") pod "ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780" (UID: "ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780"). InnerVolumeSpecName "kube-api-access-tzwkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.326208 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780" (UID: "ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.327332 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780" (UID: "ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.330750 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-config" (OuterVolumeSpecName: "config") pod "ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780" (UID: "ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.349481 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780" (UID: "ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.356522 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.356561 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.356571 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.356582 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzwkk\" (UniqueName: \"kubernetes.io/projected/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-kube-api-access-tzwkk\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.356595 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.364058 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780" (UID: "ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:12:27 crc kubenswrapper[4824]: I1209 10:12:27.459083 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.068354 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5855dddcbd-6qlwr" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.169969 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-867bb6d64d-9slcv" event={"ID":"4dc240ff-1bf1-464e-8733-59d0ae2fc4d2","Type":"ContainerStarted","Data":"74bef85f2075e623c847727b5e07e4be3267615c1d048537e2ea8e929fdcf66d"} Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.170261 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.173253 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" event={"ID":"ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780","Type":"ContainerDied","Data":"ca4d28649370aeb222a8941f8d1af79178a3523d1d6245363a4771dde7373742"} Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.173315 4824 scope.go:117] "RemoveContainer" containerID="20a51d23f331fb43fd5bb76eccbc3e81fdb36fcc151c9f8fb0f57733abeca4bf" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.173476 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-pmxwc" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.183952 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-config\") pod \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\" (UID: \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\") " Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.184048 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-combined-ca-bundle\") pod \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\" (UID: \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\") " Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.184073 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-httpd-config\") pod \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\" (UID: \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\") " Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.184255 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-ovndb-tls-certs\") pod \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\" (UID: \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\") " Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.184289 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgxh9\" (UniqueName: \"kubernetes.io/projected/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-kube-api-access-pgxh9\") pod \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\" (UID: \"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3\") " Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.186467 4824 generic.go:334] "Generic (PLEG): container finished" podID="95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3" containerID="39ac59a1f6fbf896b4c2d08f9997e40dec30a053bbbcc01bf095c383bce4b7af" exitCode=0 Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.186584 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5855dddcbd-6qlwr" event={"ID":"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3","Type":"ContainerDied","Data":"39ac59a1f6fbf896b4c2d08f9997e40dec30a053bbbcc01bf095c383bce4b7af"} Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.186644 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5855dddcbd-6qlwr" event={"ID":"95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3","Type":"ContainerDied","Data":"06ca3a68ff07de58d1ef52781bf7760d4413ace0eb94f551b127a233be609488"} Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.187051 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5855dddcbd-6qlwr" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.194591 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3" (UID: "95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.208251 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-kube-api-access-pgxh9" (OuterVolumeSpecName: "kube-api-access-pgxh9") pod "95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3" (UID: "95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3"). InnerVolumeSpecName "kube-api-access-pgxh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.253841 4824 scope.go:117] "RemoveContainer" containerID="fa297085f99c182ce1a59f3a61473b47e2dccf05183db5580ccc4e987dfa27be" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.254099 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"232c4238-3f59-4e05-ac75-8746ab0a0069","Type":"ContainerStarted","Data":"58ee5b1593c2eba87a91edb563b21a5be452ec2579087a8a1f5976bc9be26c28"} Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.254293 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="232c4238-3f59-4e05-ac75-8746ab0a0069" containerName="ceilometer-notification-agent" containerID="cri-o://89110daf7562ee61890832514845441d50bf7e08943934df95b78cd71224c749" gracePeriod=30 Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.254523 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="232c4238-3f59-4e05-ac75-8746ab0a0069" containerName="proxy-httpd" containerID="cri-o://58ee5b1593c2eba87a91edb563b21a5be452ec2579087a8a1f5976bc9be26c28" gracePeriod=30 Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.254553 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.254589 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="232c4238-3f59-4e05-ac75-8746ab0a0069" containerName="sg-core" containerID="cri-o://332f494a5454b4367146a6f5595115254e0066e1d68d691290d0464175c57efe" gracePeriod=30 Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.308491 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgxh9\" (UniqueName: \"kubernetes.io/projected/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-kube-api-access-pgxh9\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.326277 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.363882 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-867bb6d64d-9slcv" podStartSLOduration=15.363855336 podStartE2EDuration="15.363855336s" podCreationTimestamp="2025-12-09 10:12:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:12:28.232815644 +0000 UTC m=+1504.567320321" watchObservedRunningTime="2025-12-09 10:12:28.363855336 +0000 UTC m=+1504.698360003" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.385741 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-pmxwc"] Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.397030 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-pmxwc"] Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.414964 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-config" (OuterVolumeSpecName: "config") pod "95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3" (UID: "95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.430539 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.432564 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.448005 4824 scope.go:117] "RemoveContainer" containerID="8e6ba1b57bd352f45b14b90495296a5ad4334e5ba289f62bcfb6b62fdd0b9de2" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.560390 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3" (UID: "95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.560482 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 10:12:28 crc kubenswrapper[4824]: E1209 10:12:28.561205 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780" containerName="dnsmasq-dns" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.561233 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780" containerName="dnsmasq-dns" Dec 09 10:12:28 crc kubenswrapper[4824]: E1209 10:12:28.561260 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3" containerName="neutron-httpd" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.561269 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3" containerName="neutron-httpd" Dec 09 10:12:28 crc kubenswrapper[4824]: E1209 10:12:28.561290 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780" containerName="init" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.561296 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780" containerName="init" Dec 09 10:12:28 crc kubenswrapper[4824]: E1209 10:12:28.561315 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbceafb7-272a-45fc-9b56-df631986f90a" containerName="cinder-db-sync" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.561321 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbceafb7-272a-45fc-9b56-df631986f90a" containerName="cinder-db-sync" Dec 09 10:12:28 crc kubenswrapper[4824]: E1209 10:12:28.561346 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="405d0a5e-fd62-4146-bfb9-96ff6c077836" containerName="heat-db-sync" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.561352 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="405d0a5e-fd62-4146-bfb9-96ff6c077836" containerName="heat-db-sync" Dec 09 10:12:28 crc kubenswrapper[4824]: E1209 10:12:28.561372 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3" containerName="neutron-api" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.561379 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3" containerName="neutron-api" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.562300 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780" containerName="dnsmasq-dns" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.562361 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3" containerName="neutron-httpd" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.562381 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3" containerName="neutron-api" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.562393 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="405d0a5e-fd62-4146-bfb9-96ff6c077836" containerName="heat-db-sync" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.562405 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbceafb7-272a-45fc-9b56-df631986f90a" containerName="cinder-db-sync" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.564536 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.574752 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.582502 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-jgv7h"] Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.585220 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.585846 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-nzl6v" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.586205 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.588358 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.605041 4824 scope.go:117] "RemoveContainer" containerID="39ac59a1f6fbf896b4c2d08f9997e40dec30a053bbbcc01bf095c383bce4b7af" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.641885 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-jgv7h"] Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.645288 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rpxx\" (UniqueName: \"kubernetes.io/projected/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-kube-api-access-8rpxx\") pod \"cinder-scheduler-0\" (UID: \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.645335 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.645360 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-config-data\") pod \"cinder-scheduler-0\" (UID: \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.645385 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-scripts\") pod \"cinder-scheduler-0\" (UID: \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.645427 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.645455 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.645565 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.651202 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3" (UID: "95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.666435 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.679880 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.683624 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.689276 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.700652 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.748475 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-jgv7h\" (UID: \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\") " pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.748564 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rpxx\" (UniqueName: \"kubernetes.io/projected/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-kube-api-access-8rpxx\") pod \"cinder-scheduler-0\" (UID: \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.748616 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-jgv7h\" (UID: \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\") " pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.748659 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.748693 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-config-data\") pod \"cinder-scheduler-0\" (UID: \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.748739 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-jgv7h\" (UID: \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\") " pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.748801 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-scripts\") pod \"cinder-scheduler-0\" (UID: \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.748850 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-jgv7h\" (UID: \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\") " pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.752638 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.752740 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.752881 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cggz\" (UniqueName: \"kubernetes.io/projected/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-kube-api-access-7cggz\") pod \"dnsmasq-dns-5c9776ccc5-jgv7h\" (UID: \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\") " pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.753062 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-config\") pod \"dnsmasq-dns-5c9776ccc5-jgv7h\" (UID: \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\") " pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.753686 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.753719 4824 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.759912 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.760362 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-scripts\") pod \"cinder-scheduler-0\" (UID: \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.760927 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-config-data\") pod \"cinder-scheduler-0\" (UID: \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.767552 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.770288 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rpxx\" (UniqueName: \"kubernetes.io/projected/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-kube-api-access-8rpxx\") pod \"cinder-scheduler-0\" (UID: \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.781041 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.805071 4824 scope.go:117] "RemoveContainer" containerID="8e6ba1b57bd352f45b14b90495296a5ad4334e5ba289f62bcfb6b62fdd0b9de2" Dec 09 10:12:28 crc kubenswrapper[4824]: E1209 10:12:28.805889 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e6ba1b57bd352f45b14b90495296a5ad4334e5ba289f62bcfb6b62fdd0b9de2\": container with ID starting with 8e6ba1b57bd352f45b14b90495296a5ad4334e5ba289f62bcfb6b62fdd0b9de2 not found: ID does not exist" containerID="8e6ba1b57bd352f45b14b90495296a5ad4334e5ba289f62bcfb6b62fdd0b9de2" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.805939 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e6ba1b57bd352f45b14b90495296a5ad4334e5ba289f62bcfb6b62fdd0b9de2"} err="failed to get container status \"8e6ba1b57bd352f45b14b90495296a5ad4334e5ba289f62bcfb6b62fdd0b9de2\": rpc error: code = NotFound desc = could not find container \"8e6ba1b57bd352f45b14b90495296a5ad4334e5ba289f62bcfb6b62fdd0b9de2\": container with ID starting with 8e6ba1b57bd352f45b14b90495296a5ad4334e5ba289f62bcfb6b62fdd0b9de2 not found: ID does not exist" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.805975 4824 scope.go:117] "RemoveContainer" containerID="39ac59a1f6fbf896b4c2d08f9997e40dec30a053bbbcc01bf095c383bce4b7af" Dec 09 10:12:28 crc kubenswrapper[4824]: E1209 10:12:28.807146 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39ac59a1f6fbf896b4c2d08f9997e40dec30a053bbbcc01bf095c383bce4b7af\": container with ID starting with 39ac59a1f6fbf896b4c2d08f9997e40dec30a053bbbcc01bf095c383bce4b7af not found: ID does not exist" containerID="39ac59a1f6fbf896b4c2d08f9997e40dec30a053bbbcc01bf095c383bce4b7af" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.807194 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39ac59a1f6fbf896b4c2d08f9997e40dec30a053bbbcc01bf095c383bce4b7af"} err="failed to get container status \"39ac59a1f6fbf896b4c2d08f9997e40dec30a053bbbcc01bf095c383bce4b7af\": rpc error: code = NotFound desc = could not find container \"39ac59a1f6fbf896b4c2d08f9997e40dec30a053bbbcc01bf095c383bce4b7af\": container with ID starting with 39ac59a1f6fbf896b4c2d08f9997e40dec30a053bbbcc01bf095c383bce4b7af not found: ID does not exist" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.858577 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-scripts\") pod \"cinder-api-0\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " pod="openstack/cinder-api-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.858716 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-jgv7h\" (UID: \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\") " pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.858817 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-jgv7h\" (UID: \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\") " pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.858906 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-jgv7h\" (UID: \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\") " pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.858991 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-config-data\") pod \"cinder-api-0\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " pod="openstack/cinder-api-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.859059 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cggz\" (UniqueName: \"kubernetes.io/projected/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-kube-api-access-7cggz\") pod \"dnsmasq-dns-5c9776ccc5-jgv7h\" (UID: \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\") " pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.859097 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcf5b\" (UniqueName: \"kubernetes.io/projected/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-kube-api-access-gcf5b\") pod \"cinder-api-0\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " pod="openstack/cinder-api-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.859146 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " pod="openstack/cinder-api-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.859212 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-config\") pod \"dnsmasq-dns-5c9776ccc5-jgv7h\" (UID: \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\") " pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.859299 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-logs\") pod \"cinder-api-0\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " pod="openstack/cinder-api-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.859345 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " pod="openstack/cinder-api-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.859373 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-config-data-custom\") pod \"cinder-api-0\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " pod="openstack/cinder-api-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.859398 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-jgv7h\" (UID: \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\") " pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.860226 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-jgv7h\" (UID: \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\") " pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.860512 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-jgv7h\" (UID: \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\") " pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.861500 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-config\") pod \"dnsmasq-dns-5c9776ccc5-jgv7h\" (UID: \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\") " pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.862028 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-jgv7h\" (UID: \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\") " pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.862202 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-jgv7h\" (UID: \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\") " pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.866141 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5855dddcbd-6qlwr"] Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.884223 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cggz\" (UniqueName: \"kubernetes.io/projected/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-kube-api-access-7cggz\") pod \"dnsmasq-dns-5c9776ccc5-jgv7h\" (UID: \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\") " pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.886902 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5855dddcbd-6qlwr"] Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.962172 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-logs\") pod \"cinder-api-0\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " pod="openstack/cinder-api-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.962494 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " pod="openstack/cinder-api-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.962520 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-config-data-custom\") pod \"cinder-api-0\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " pod="openstack/cinder-api-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.962576 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-scripts\") pod \"cinder-api-0\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " pod="openstack/cinder-api-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.962593 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " pod="openstack/cinder-api-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.962733 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-config-data\") pod \"cinder-api-0\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " pod="openstack/cinder-api-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.962811 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcf5b\" (UniqueName: \"kubernetes.io/projected/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-kube-api-access-gcf5b\") pod \"cinder-api-0\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " pod="openstack/cinder-api-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.962844 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " pod="openstack/cinder-api-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.963399 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-logs\") pod \"cinder-api-0\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " pod="openstack/cinder-api-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.968072 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-config-data\") pod \"cinder-api-0\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " pod="openstack/cinder-api-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.971295 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-config-data-custom\") pod \"cinder-api-0\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " pod="openstack/cinder-api-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.972215 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " pod="openstack/cinder-api-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.978975 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-scripts\") pod \"cinder-api-0\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " pod="openstack/cinder-api-0" Dec 09 10:12:28 crc kubenswrapper[4824]: I1209 10:12:28.987896 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcf5b\" (UniqueName: \"kubernetes.io/projected/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-kube-api-access-gcf5b\") pod \"cinder-api-0\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " pod="openstack/cinder-api-0" Dec 09 10:12:29 crc kubenswrapper[4824]: I1209 10:12:29.106571 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" Dec 09 10:12:29 crc kubenswrapper[4824]: I1209 10:12:29.120127 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 10:12:29 crc kubenswrapper[4824]: I1209 10:12:29.316195 4824 generic.go:334] "Generic (PLEG): container finished" podID="232c4238-3f59-4e05-ac75-8746ab0a0069" containerID="58ee5b1593c2eba87a91edb563b21a5be452ec2579087a8a1f5976bc9be26c28" exitCode=0 Dec 09 10:12:29 crc kubenswrapper[4824]: I1209 10:12:29.316234 4824 generic.go:334] "Generic (PLEG): container finished" podID="232c4238-3f59-4e05-ac75-8746ab0a0069" containerID="332f494a5454b4367146a6f5595115254e0066e1d68d691290d0464175c57efe" exitCode=2 Dec 09 10:12:29 crc kubenswrapper[4824]: I1209 10:12:29.316301 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"232c4238-3f59-4e05-ac75-8746ab0a0069","Type":"ContainerDied","Data":"58ee5b1593c2eba87a91edb563b21a5be452ec2579087a8a1f5976bc9be26c28"} Dec 09 10:12:29 crc kubenswrapper[4824]: I1209 10:12:29.316327 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"232c4238-3f59-4e05-ac75-8746ab0a0069","Type":"ContainerDied","Data":"332f494a5454b4367146a6f5595115254e0066e1d68d691290d0464175c57efe"} Dec 09 10:12:29 crc kubenswrapper[4824]: I1209 10:12:29.451665 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lhlfr" podUID="ad8cfdb4-f7e5-4b30-a1b8-a680b111f788" containerName="registry-server" probeResult="failure" output=< Dec 09 10:12:29 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 10:12:29 crc kubenswrapper[4824]: > Dec 09 10:12:29 crc kubenswrapper[4824]: I1209 10:12:29.562349 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 10:12:29 crc kubenswrapper[4824]: I1209 10:12:29.615721 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 10:12:29 crc kubenswrapper[4824]: I1209 10:12:29.903544 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 09 10:12:29 crc kubenswrapper[4824]: I1209 10:12:29.964186 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3" path="/var/lib/kubelet/pods/95a62eb4-e5ed-4e2d-adef-7bb12f4f77d3/volumes" Dec 09 10:12:29 crc kubenswrapper[4824]: I1209 10:12:29.965004 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780" path="/var/lib/kubelet/pods/ecbd18ee-1433-4ad0-8ab2-3ad4d1f9f780/volumes" Dec 09 10:12:30 crc kubenswrapper[4824]: I1209 10:12:30.191696 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-jgv7h"] Dec 09 10:12:30 crc kubenswrapper[4824]: I1209 10:12:30.371167 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b","Type":"ContainerStarted","Data":"ea28c43ec2abec5438ddaa9acd743f758507ffd3669148871909fd04484e3cf0"} Dec 09 10:12:30 crc kubenswrapper[4824]: I1209 10:12:30.375169 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5","Type":"ContainerStarted","Data":"a81cdeb105fa88f95542073b2f7f749454390584714885ceb38ec6192c185f46"} Dec 09 10:12:30 crc kubenswrapper[4824]: I1209 10:12:30.378215 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 10:12:30 crc kubenswrapper[4824]: I1209 10:12:30.378201 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" event={"ID":"6b816dc5-829b-4c38-b2e0-b08cfbfc1617","Type":"ContainerStarted","Data":"85a0480dda0da6396564457bad6265b5bbf437b535fb5c1209c8ab3781aff40a"} Dec 09 10:12:30 crc kubenswrapper[4824]: I1209 10:12:30.783157 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 09 10:12:31 crc kubenswrapper[4824]: I1209 10:12:31.409579 4824 generic.go:334] "Generic (PLEG): container finished" podID="6b816dc5-829b-4c38-b2e0-b08cfbfc1617" containerID="114b030ac8493b55488b9a7a828053a7e865af0e63128f8d2ffb9d91a3d972f7" exitCode=0 Dec 09 10:12:31 crc kubenswrapper[4824]: I1209 10:12:31.410102 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" event={"ID":"6b816dc5-829b-4c38-b2e0-b08cfbfc1617","Type":"ContainerDied","Data":"114b030ac8493b55488b9a7a828053a7e865af0e63128f8d2ffb9d91a3d972f7"} Dec 09 10:12:31 crc kubenswrapper[4824]: I1209 10:12:31.422478 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b","Type":"ContainerStarted","Data":"d4512841b2875155ab8b8e4cd53333f67457e28ae214e9905332d1538133d993"} Dec 09 10:12:31 crc kubenswrapper[4824]: I1209 10:12:31.430493 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"232c4238-3f59-4e05-ac75-8746ab0a0069","Type":"ContainerDied","Data":"89110daf7562ee61890832514845441d50bf7e08943934df95b78cd71224c749"} Dec 09 10:12:31 crc kubenswrapper[4824]: I1209 10:12:31.430440 4824 generic.go:334] "Generic (PLEG): container finished" podID="232c4238-3f59-4e05-ac75-8746ab0a0069" containerID="89110daf7562ee61890832514845441d50bf7e08943934df95b78cd71224c749" exitCode=0 Dec 09 10:12:31 crc kubenswrapper[4824]: I1209 10:12:31.974151 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.031714 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.141986 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/232c4238-3f59-4e05-ac75-8746ab0a0069-scripts\") pod \"232c4238-3f59-4e05-ac75-8746ab0a0069\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.142081 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/232c4238-3f59-4e05-ac75-8746ab0a0069-run-httpd\") pod \"232c4238-3f59-4e05-ac75-8746ab0a0069\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.142127 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/232c4238-3f59-4e05-ac75-8746ab0a0069-config-data\") pod \"232c4238-3f59-4e05-ac75-8746ab0a0069\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.142215 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/232c4238-3f59-4e05-ac75-8746ab0a0069-log-httpd\") pod \"232c4238-3f59-4e05-ac75-8746ab0a0069\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.142327 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/232c4238-3f59-4e05-ac75-8746ab0a0069-sg-core-conf-yaml\") pod \"232c4238-3f59-4e05-ac75-8746ab0a0069\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.142369 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/232c4238-3f59-4e05-ac75-8746ab0a0069-combined-ca-bundle\") pod \"232c4238-3f59-4e05-ac75-8746ab0a0069\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.142438 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6zsx\" (UniqueName: \"kubernetes.io/projected/232c4238-3f59-4e05-ac75-8746ab0a0069-kube-api-access-r6zsx\") pod \"232c4238-3f59-4e05-ac75-8746ab0a0069\" (UID: \"232c4238-3f59-4e05-ac75-8746ab0a0069\") " Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.144674 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/232c4238-3f59-4e05-ac75-8746ab0a0069-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "232c4238-3f59-4e05-ac75-8746ab0a0069" (UID: "232c4238-3f59-4e05-ac75-8746ab0a0069"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.144966 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/232c4238-3f59-4e05-ac75-8746ab0a0069-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "232c4238-3f59-4e05-ac75-8746ab0a0069" (UID: "232c4238-3f59-4e05-ac75-8746ab0a0069"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.153217 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/232c4238-3f59-4e05-ac75-8746ab0a0069-kube-api-access-r6zsx" (OuterVolumeSpecName: "kube-api-access-r6zsx") pod "232c4238-3f59-4e05-ac75-8746ab0a0069" (UID: "232c4238-3f59-4e05-ac75-8746ab0a0069"). InnerVolumeSpecName "kube-api-access-r6zsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.168348 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/232c4238-3f59-4e05-ac75-8746ab0a0069-scripts" (OuterVolumeSpecName: "scripts") pod "232c4238-3f59-4e05-ac75-8746ab0a0069" (UID: "232c4238-3f59-4e05-ac75-8746ab0a0069"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.202512 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/232c4238-3f59-4e05-ac75-8746ab0a0069-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "232c4238-3f59-4e05-ac75-8746ab0a0069" (UID: "232c4238-3f59-4e05-ac75-8746ab0a0069"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.247797 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/232c4238-3f59-4e05-ac75-8746ab0a0069-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.247838 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/232c4238-3f59-4e05-ac75-8746ab0a0069-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.247850 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/232c4238-3f59-4e05-ac75-8746ab0a0069-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.247860 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/232c4238-3f59-4e05-ac75-8746ab0a0069-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.247872 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6zsx\" (UniqueName: \"kubernetes.io/projected/232c4238-3f59-4e05-ac75-8746ab0a0069-kube-api-access-r6zsx\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.256439 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/232c4238-3f59-4e05-ac75-8746ab0a0069-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "232c4238-3f59-4e05-ac75-8746ab0a0069" (UID: "232c4238-3f59-4e05-ac75-8746ab0a0069"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.299564 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/232c4238-3f59-4e05-ac75-8746ab0a0069-config-data" (OuterVolumeSpecName: "config-data") pod "232c4238-3f59-4e05-ac75-8746ab0a0069" (UID: "232c4238-3f59-4e05-ac75-8746ab0a0069"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.350008 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/232c4238-3f59-4e05-ac75-8746ab0a0069-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.350055 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/232c4238-3f59-4e05-ac75-8746ab0a0069-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.456935 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" event={"ID":"6b816dc5-829b-4c38-b2e0-b08cfbfc1617","Type":"ContainerStarted","Data":"29c9a683ae0d4909abd63d7f40848eeabda0f73ae8dea7e1587f0586061f08a4"} Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.457568 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.462343 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b","Type":"ContainerStarted","Data":"09192e79e2b24b4ba4ff045af664c1a16f27e0e694d1b16486716f09b6c84652"} Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.462453 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.462456 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b" containerName="cinder-api-log" containerID="cri-o://d4512841b2875155ab8b8e4cd53333f67457e28ae214e9905332d1538133d993" gracePeriod=30 Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.462506 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b" containerName="cinder-api" containerID="cri-o://09192e79e2b24b4ba4ff045af664c1a16f27e0e694d1b16486716f09b6c84652" gracePeriod=30 Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.465920 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5","Type":"ContainerStarted","Data":"9a7e126bfed9f9eccf9686b200bfbceca110dcae02d5023b921a7b86b96e0d92"} Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.480706 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" podStartSLOduration=4.480672461 podStartE2EDuration="4.480672461s" podCreationTimestamp="2025-12-09 10:12:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:12:32.475849257 +0000 UTC m=+1508.810353934" watchObservedRunningTime="2025-12-09 10:12:32.480672461 +0000 UTC m=+1508.815177128" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.483442 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"232c4238-3f59-4e05-ac75-8746ab0a0069","Type":"ContainerDied","Data":"ed62ce1026e7eb19f4d8be85136efcd90bca262ff995fe62cb129a67e7e968de"} Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.483496 4824 scope.go:117] "RemoveContainer" containerID="58ee5b1593c2eba87a91edb563b21a5be452ec2579087a8a1f5976bc9be26c28" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.483704 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.504514 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.504462678 podStartE2EDuration="4.504462678s" podCreationTimestamp="2025-12-09 10:12:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:12:32.502627749 +0000 UTC m=+1508.837132436" watchObservedRunningTime="2025-12-09 10:12:32.504462678 +0000 UTC m=+1508.838967345" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.549617 4824 scope.go:117] "RemoveContainer" containerID="332f494a5454b4367146a6f5595115254e0066e1d68d691290d0464175c57efe" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.613689 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.633346 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.660916 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:12:32 crc kubenswrapper[4824]: E1209 10:12:32.662663 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="232c4238-3f59-4e05-ac75-8746ab0a0069" containerName="proxy-httpd" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.663377 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="232c4238-3f59-4e05-ac75-8746ab0a0069" containerName="proxy-httpd" Dec 09 10:12:32 crc kubenswrapper[4824]: E1209 10:12:32.663418 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="232c4238-3f59-4e05-ac75-8746ab0a0069" containerName="ceilometer-notification-agent" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.663428 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="232c4238-3f59-4e05-ac75-8746ab0a0069" containerName="ceilometer-notification-agent" Dec 09 10:12:32 crc kubenswrapper[4824]: E1209 10:12:32.663443 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="232c4238-3f59-4e05-ac75-8746ab0a0069" containerName="sg-core" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.663450 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="232c4238-3f59-4e05-ac75-8746ab0a0069" containerName="sg-core" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.663798 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="232c4238-3f59-4e05-ac75-8746ab0a0069" containerName="proxy-httpd" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.663827 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="232c4238-3f59-4e05-ac75-8746ab0a0069" containerName="sg-core" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.663861 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="232c4238-3f59-4e05-ac75-8746ab0a0069" containerName="ceilometer-notification-agent" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.667305 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.670848 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.671050 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.706147 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.711197 4824 scope.go:117] "RemoveContainer" containerID="89110daf7562ee61890832514845441d50bf7e08943934df95b78cd71224c749" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.792941 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188ccefd-703d-4857-a005-70c0ce68d09c-config-data\") pod \"ceilometer-0\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " pod="openstack/ceilometer-0" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.793003 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/188ccefd-703d-4857-a005-70c0ce68d09c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " pod="openstack/ceilometer-0" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.794442 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/188ccefd-703d-4857-a005-70c0ce68d09c-run-httpd\") pod \"ceilometer-0\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " pod="openstack/ceilometer-0" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.794474 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188ccefd-703d-4857-a005-70c0ce68d09c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " pod="openstack/ceilometer-0" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.794658 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/188ccefd-703d-4857-a005-70c0ce68d09c-scripts\") pod \"ceilometer-0\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " pod="openstack/ceilometer-0" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.794808 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7k6l\" (UniqueName: \"kubernetes.io/projected/188ccefd-703d-4857-a005-70c0ce68d09c-kube-api-access-p7k6l\") pod \"ceilometer-0\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " pod="openstack/ceilometer-0" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.794840 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/188ccefd-703d-4857-a005-70c0ce68d09c-log-httpd\") pod \"ceilometer-0\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " pod="openstack/ceilometer-0" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.896711 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/188ccefd-703d-4857-a005-70c0ce68d09c-scripts\") pod \"ceilometer-0\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " pod="openstack/ceilometer-0" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.897281 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7k6l\" (UniqueName: \"kubernetes.io/projected/188ccefd-703d-4857-a005-70c0ce68d09c-kube-api-access-p7k6l\") pod \"ceilometer-0\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " pod="openstack/ceilometer-0" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.897384 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/188ccefd-703d-4857-a005-70c0ce68d09c-log-httpd\") pod \"ceilometer-0\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " pod="openstack/ceilometer-0" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.897590 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188ccefd-703d-4857-a005-70c0ce68d09c-config-data\") pod \"ceilometer-0\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " pod="openstack/ceilometer-0" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.897708 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/188ccefd-703d-4857-a005-70c0ce68d09c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " pod="openstack/ceilometer-0" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.897896 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/188ccefd-703d-4857-a005-70c0ce68d09c-run-httpd\") pod \"ceilometer-0\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " pod="openstack/ceilometer-0" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.898000 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188ccefd-703d-4857-a005-70c0ce68d09c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " pod="openstack/ceilometer-0" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.898215 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/188ccefd-703d-4857-a005-70c0ce68d09c-run-httpd\") pod \"ceilometer-0\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " pod="openstack/ceilometer-0" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.897931 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/188ccefd-703d-4857-a005-70c0ce68d09c-log-httpd\") pod \"ceilometer-0\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " pod="openstack/ceilometer-0" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.909857 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/188ccefd-703d-4857-a005-70c0ce68d09c-scripts\") pod \"ceilometer-0\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " pod="openstack/ceilometer-0" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.916126 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188ccefd-703d-4857-a005-70c0ce68d09c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " pod="openstack/ceilometer-0" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.916264 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188ccefd-703d-4857-a005-70c0ce68d09c-config-data\") pod \"ceilometer-0\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " pod="openstack/ceilometer-0" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.931733 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7k6l\" (UniqueName: \"kubernetes.io/projected/188ccefd-703d-4857-a005-70c0ce68d09c-kube-api-access-p7k6l\") pod \"ceilometer-0\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " pod="openstack/ceilometer-0" Dec 09 10:12:32 crc kubenswrapper[4824]: I1209 10:12:32.933213 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/188ccefd-703d-4857-a005-70c0ce68d09c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " pod="openstack/ceilometer-0" Dec 09 10:12:33 crc kubenswrapper[4824]: I1209 10:12:33.035524 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:12:33 crc kubenswrapper[4824]: I1209 10:12:33.566936 4824 generic.go:334] "Generic (PLEG): container finished" podID="6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b" containerID="d4512841b2875155ab8b8e4cd53333f67457e28ae214e9905332d1538133d993" exitCode=143 Dec 09 10:12:33 crc kubenswrapper[4824]: I1209 10:12:33.567042 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b","Type":"ContainerDied","Data":"d4512841b2875155ab8b8e4cd53333f67457e28ae214e9905332d1538133d993"} Dec 09 10:12:33 crc kubenswrapper[4824]: I1209 10:12:33.583595 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5","Type":"ContainerStarted","Data":"355cf99d0ef56866339c2aca549bfc6cb1fed56cd38dd39873d7fefed6d6e046"} Dec 09 10:12:33 crc kubenswrapper[4824]: I1209 10:12:33.711155 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.637777113 podStartE2EDuration="5.711127166s" podCreationTimestamp="2025-12-09 10:12:28 +0000 UTC" firstStartedPulling="2025-12-09 10:12:29.615390841 +0000 UTC m=+1505.949895508" lastFinishedPulling="2025-12-09 10:12:30.688740894 +0000 UTC m=+1507.023245561" observedRunningTime="2025-12-09 10:12:33.690488418 +0000 UTC m=+1510.024993095" watchObservedRunningTime="2025-12-09 10:12:33.711127166 +0000 UTC m=+1510.045631853" Dec 09 10:12:33 crc kubenswrapper[4824]: I1209 10:12:33.780610 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:12:33 crc kubenswrapper[4824]: I1209 10:12:33.789948 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 09 10:12:33 crc kubenswrapper[4824]: I1209 10:12:33.940848 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="232c4238-3f59-4e05-ac75-8746ab0a0069" path="/var/lib/kubelet/pods/232c4238-3f59-4e05-ac75-8746ab0a0069/volumes" Dec 09 10:12:34 crc kubenswrapper[4824]: I1209 10:12:34.636388 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"188ccefd-703d-4857-a005-70c0ce68d09c","Type":"ContainerStarted","Data":"86c269b7bff445ae59c766a934fd866cc747ab648a33b2f9c4e8ec19aef44a93"} Dec 09 10:12:34 crc kubenswrapper[4824]: I1209 10:12:34.636879 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"188ccefd-703d-4857-a005-70c0ce68d09c","Type":"ContainerStarted","Data":"619d823c4dcabead9a5845864f1822ecce507d47502f5efe978f61491f5d3bfb"} Dec 09 10:12:35 crc kubenswrapper[4824]: I1209 10:12:35.566280 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-867bb6d64d-9slcv" Dec 09 10:12:35 crc kubenswrapper[4824]: I1209 10:12:35.688124 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"188ccefd-703d-4857-a005-70c0ce68d09c","Type":"ContainerStarted","Data":"fe26ca2e87cd228b8c6a6502631ef82a32112cc0374160767f4acd98b820dd7b"} Dec 09 10:12:35 crc kubenswrapper[4824]: I1209 10:12:35.694965 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5f9b44b4db-xxc6m"] Dec 09 10:12:35 crc kubenswrapper[4824]: I1209 10:12:35.695311 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5f9b44b4db-xxc6m" podUID="9f8a567c-2577-4c66-a75e-e43f7cd66000" containerName="barbican-api-log" containerID="cri-o://b5cf173a26f1977cf61d3b4184fa796463dc2a63ac661f657cca3f3e873af83f" gracePeriod=30 Dec 09 10:12:35 crc kubenswrapper[4824]: I1209 10:12:35.695508 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5f9b44b4db-xxc6m" podUID="9f8a567c-2577-4c66-a75e-e43f7cd66000" containerName="barbican-api" containerID="cri-o://5ea3d7e5055f26b89248f5118b0972582d7d96bea9dedca1d4da15bbb311e5d7" gracePeriod=30 Dec 09 10:12:36 crc kubenswrapper[4824]: I1209 10:12:36.701167 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"188ccefd-703d-4857-a005-70c0ce68d09c","Type":"ContainerStarted","Data":"060ba5d4bcaabe29261b98d0afd9b5e28d1447b92d0016faa1fabc5823425c22"} Dec 09 10:12:36 crc kubenswrapper[4824]: I1209 10:12:36.703709 4824 generic.go:334] "Generic (PLEG): container finished" podID="9f8a567c-2577-4c66-a75e-e43f7cd66000" containerID="b5cf173a26f1977cf61d3b4184fa796463dc2a63ac661f657cca3f3e873af83f" exitCode=143 Dec 09 10:12:36 crc kubenswrapper[4824]: I1209 10:12:36.703817 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f9b44b4db-xxc6m" event={"ID":"9f8a567c-2577-4c66-a75e-e43f7cd66000","Type":"ContainerDied","Data":"b5cf173a26f1977cf61d3b4184fa796463dc2a63ac661f657cca3f3e873af83f"} Dec 09 10:12:37 crc kubenswrapper[4824]: I1209 10:12:37.620701 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wx2h9"] Dec 09 10:12:37 crc kubenswrapper[4824]: I1209 10:12:37.639286 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wx2h9" Dec 09 10:12:37 crc kubenswrapper[4824]: I1209 10:12:37.690269 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wx2h9"] Dec 09 10:12:37 crc kubenswrapper[4824]: I1209 10:12:37.805210 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/065686b2-246c-4194-936f-b6e58ee9763b-utilities\") pod \"community-operators-wx2h9\" (UID: \"065686b2-246c-4194-936f-b6e58ee9763b\") " pod="openshift-marketplace/community-operators-wx2h9" Dec 09 10:12:37 crc kubenswrapper[4824]: I1209 10:12:37.805720 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/065686b2-246c-4194-936f-b6e58ee9763b-catalog-content\") pod \"community-operators-wx2h9\" (UID: \"065686b2-246c-4194-936f-b6e58ee9763b\") " pod="openshift-marketplace/community-operators-wx2h9" Dec 09 10:12:37 crc kubenswrapper[4824]: I1209 10:12:37.805890 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwfmq\" (UniqueName: \"kubernetes.io/projected/065686b2-246c-4194-936f-b6e58ee9763b-kube-api-access-bwfmq\") pod \"community-operators-wx2h9\" (UID: \"065686b2-246c-4194-936f-b6e58ee9763b\") " pod="openshift-marketplace/community-operators-wx2h9" Dec 09 10:12:37 crc kubenswrapper[4824]: I1209 10:12:37.908242 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/065686b2-246c-4194-936f-b6e58ee9763b-utilities\") pod \"community-operators-wx2h9\" (UID: \"065686b2-246c-4194-936f-b6e58ee9763b\") " pod="openshift-marketplace/community-operators-wx2h9" Dec 09 10:12:37 crc kubenswrapper[4824]: I1209 10:12:37.908331 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/065686b2-246c-4194-936f-b6e58ee9763b-catalog-content\") pod \"community-operators-wx2h9\" (UID: \"065686b2-246c-4194-936f-b6e58ee9763b\") " pod="openshift-marketplace/community-operators-wx2h9" Dec 09 10:12:37 crc kubenswrapper[4824]: I1209 10:12:37.908422 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwfmq\" (UniqueName: \"kubernetes.io/projected/065686b2-246c-4194-936f-b6e58ee9763b-kube-api-access-bwfmq\") pod \"community-operators-wx2h9\" (UID: \"065686b2-246c-4194-936f-b6e58ee9763b\") " pod="openshift-marketplace/community-operators-wx2h9" Dec 09 10:12:37 crc kubenswrapper[4824]: I1209 10:12:37.908928 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/065686b2-246c-4194-936f-b6e58ee9763b-utilities\") pod \"community-operators-wx2h9\" (UID: \"065686b2-246c-4194-936f-b6e58ee9763b\") " pod="openshift-marketplace/community-operators-wx2h9" Dec 09 10:12:37 crc kubenswrapper[4824]: I1209 10:12:37.909254 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/065686b2-246c-4194-936f-b6e58ee9763b-catalog-content\") pod \"community-operators-wx2h9\" (UID: \"065686b2-246c-4194-936f-b6e58ee9763b\") " pod="openshift-marketplace/community-operators-wx2h9" Dec 09 10:12:37 crc kubenswrapper[4824]: I1209 10:12:37.926422 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwfmq\" (UniqueName: \"kubernetes.io/projected/065686b2-246c-4194-936f-b6e58ee9763b-kube-api-access-bwfmq\") pod \"community-operators-wx2h9\" (UID: \"065686b2-246c-4194-936f-b6e58ee9763b\") " pod="openshift-marketplace/community-operators-wx2h9" Dec 09 10:12:38 crc kubenswrapper[4824]: I1209 10:12:38.034706 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wx2h9" Dec 09 10:12:38 crc kubenswrapper[4824]: I1209 10:12:38.303320 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lhlfr" Dec 09 10:12:38 crc kubenswrapper[4824]: I1209 10:12:38.365726 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lhlfr" Dec 09 10:12:38 crc kubenswrapper[4824]: W1209 10:12:38.608976 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod065686b2_246c_4194_936f_b6e58ee9763b.slice/crio-99ff906e3a59ca7f072add5e8e008da8d591b81df699822444eebd0ce7743340 WatchSource:0}: Error finding container 99ff906e3a59ca7f072add5e8e008da8d591b81df699822444eebd0ce7743340: Status 404 returned error can't find the container with id 99ff906e3a59ca7f072add5e8e008da8d591b81df699822444eebd0ce7743340 Dec 09 10:12:38 crc kubenswrapper[4824]: I1209 10:12:38.615220 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wx2h9"] Dec 09 10:12:38 crc kubenswrapper[4824]: I1209 10:12:38.730729 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wx2h9" event={"ID":"065686b2-246c-4194-936f-b6e58ee9763b","Type":"ContainerStarted","Data":"99ff906e3a59ca7f072add5e8e008da8d591b81df699822444eebd0ce7743340"} Dec 09 10:12:38 crc kubenswrapper[4824]: I1209 10:12:38.734295 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"188ccefd-703d-4857-a005-70c0ce68d09c","Type":"ContainerStarted","Data":"dbc7f9130eac008823602e68826e11f4cd01648aa902f777ec75e36f8d466412"} Dec 09 10:12:38 crc kubenswrapper[4824]: I1209 10:12:38.766456 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.01633131 podStartE2EDuration="6.766429423s" podCreationTimestamp="2025-12-09 10:12:32 +0000 UTC" firstStartedPulling="2025-12-09 10:12:33.750078365 +0000 UTC m=+1510.084583032" lastFinishedPulling="2025-12-09 10:12:37.500176478 +0000 UTC m=+1513.834681145" observedRunningTime="2025-12-09 10:12:38.754600596 +0000 UTC m=+1515.089105263" watchObservedRunningTime="2025-12-09 10:12:38.766429423 +0000 UTC m=+1515.100934100" Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.058014 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.110813 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.129076 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.200473 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-2952m"] Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.200839 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85ff748b95-2952m" podUID="26b61717-1239-4782-bc75-6b0eece01c14" containerName="dnsmasq-dns" containerID="cri-o://bc738525e61a76ad921d11262bd0345eb8616de11b3587e7a1de703692188cda" gracePeriod=10 Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.602539 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f9b44b4db-xxc6m" Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.685868 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f8a567c-2577-4c66-a75e-e43f7cd66000-logs\") pod \"9f8a567c-2577-4c66-a75e-e43f7cd66000\" (UID: \"9f8a567c-2577-4c66-a75e-e43f7cd66000\") " Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.686090 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f8a567c-2577-4c66-a75e-e43f7cd66000-config-data\") pod \"9f8a567c-2577-4c66-a75e-e43f7cd66000\" (UID: \"9f8a567c-2577-4c66-a75e-e43f7cd66000\") " Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.686143 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f8a567c-2577-4c66-a75e-e43f7cd66000-combined-ca-bundle\") pod \"9f8a567c-2577-4c66-a75e-e43f7cd66000\" (UID: \"9f8a567c-2577-4c66-a75e-e43f7cd66000\") " Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.686187 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jp4c2\" (UniqueName: \"kubernetes.io/projected/9f8a567c-2577-4c66-a75e-e43f7cd66000-kube-api-access-jp4c2\") pod \"9f8a567c-2577-4c66-a75e-e43f7cd66000\" (UID: \"9f8a567c-2577-4c66-a75e-e43f7cd66000\") " Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.686273 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f8a567c-2577-4c66-a75e-e43f7cd66000-config-data-custom\") pod \"9f8a567c-2577-4c66-a75e-e43f7cd66000\" (UID: \"9f8a567c-2577-4c66-a75e-e43f7cd66000\") " Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.687245 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f8a567c-2577-4c66-a75e-e43f7cd66000-logs" (OuterVolumeSpecName: "logs") pod "9f8a567c-2577-4c66-a75e-e43f7cd66000" (UID: "9f8a567c-2577-4c66-a75e-e43f7cd66000"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.705161 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f8a567c-2577-4c66-a75e-e43f7cd66000-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9f8a567c-2577-4c66-a75e-e43f7cd66000" (UID: "9f8a567c-2577-4c66-a75e-e43f7cd66000"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.714149 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f8a567c-2577-4c66-a75e-e43f7cd66000-kube-api-access-jp4c2" (OuterVolumeSpecName: "kube-api-access-jp4c2") pod "9f8a567c-2577-4c66-a75e-e43f7cd66000" (UID: "9f8a567c-2577-4c66-a75e-e43f7cd66000"). InnerVolumeSpecName "kube-api-access-jp4c2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.789008 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f8a567c-2577-4c66-a75e-e43f7cd66000-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.789046 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f8a567c-2577-4c66-a75e-e43f7cd66000-logs\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.789062 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jp4c2\" (UniqueName: \"kubernetes.io/projected/9f8a567c-2577-4c66-a75e-e43f7cd66000-kube-api-access-jp4c2\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.797607 4824 generic.go:334] "Generic (PLEG): container finished" podID="9f8a567c-2577-4c66-a75e-e43f7cd66000" containerID="5ea3d7e5055f26b89248f5118b0972582d7d96bea9dedca1d4da15bbb311e5d7" exitCode=0 Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.797666 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f9b44b4db-xxc6m" event={"ID":"9f8a567c-2577-4c66-a75e-e43f7cd66000","Type":"ContainerDied","Data":"5ea3d7e5055f26b89248f5118b0972582d7d96bea9dedca1d4da15bbb311e5d7"} Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.797694 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f9b44b4db-xxc6m" event={"ID":"9f8a567c-2577-4c66-a75e-e43f7cd66000","Type":"ContainerDied","Data":"04e9bcf61ce76c0a2526dca38ba02d2faf65851ef66dbf13fd8b4a6a18f6acb8"} Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.797711 4824 scope.go:117] "RemoveContainer" containerID="5ea3d7e5055f26b89248f5118b0972582d7d96bea9dedca1d4da15bbb311e5d7" Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.797878 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f9b44b4db-xxc6m" Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.799956 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f8a567c-2577-4c66-a75e-e43f7cd66000-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f8a567c-2577-4c66-a75e-e43f7cd66000" (UID: "9f8a567c-2577-4c66-a75e-e43f7cd66000"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.833571 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b61717-1239-4782-bc75-6b0eece01c14" containerID="bc738525e61a76ad921d11262bd0345eb8616de11b3587e7a1de703692188cda" exitCode=0 Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.833668 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-2952m" event={"ID":"26b61717-1239-4782-bc75-6b0eece01c14","Type":"ContainerDied","Data":"bc738525e61a76ad921d11262bd0345eb8616de11b3587e7a1de703692188cda"} Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.842030 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f8a567c-2577-4c66-a75e-e43f7cd66000-config-data" (OuterVolumeSpecName: "config-data") pod "9f8a567c-2577-4c66-a75e-e43f7cd66000" (UID: "9f8a567c-2577-4c66-a75e-e43f7cd66000"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.842975 4824 generic.go:334] "Generic (PLEG): container finished" podID="065686b2-246c-4194-936f-b6e58ee9763b" containerID="84d3866e15cc80d2a49e7eb9f9cf31c3dfc4bdf745a44a9943a62dbd7a5e0b4b" exitCode=0 Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.843464 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wx2h9" event={"ID":"065686b2-246c-4194-936f-b6e58ee9763b","Type":"ContainerDied","Data":"84d3866e15cc80d2a49e7eb9f9cf31c3dfc4bdf745a44a9943a62dbd7a5e0b4b"} Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.843734 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5" containerName="cinder-scheduler" containerID="cri-o://9a7e126bfed9f9eccf9686b200bfbceca110dcae02d5023b921a7b86b96e0d92" gracePeriod=30 Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.844434 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5" containerName="probe" containerID="cri-o://355cf99d0ef56866339c2aca549bfc6cb1fed56cd38dd39873d7fefed6d6e046" gracePeriod=30 Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.844830 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.896105 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f8a567c-2577-4c66-a75e-e43f7cd66000-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.896138 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f8a567c-2577-4c66-a75e-e43f7cd66000-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.984679 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-2952m" Dec 09 10:12:39 crc kubenswrapper[4824]: I1209 10:12:39.993055 4824 scope.go:117] "RemoveContainer" containerID="b5cf173a26f1977cf61d3b4184fa796463dc2a63ac661f657cca3f3e873af83f" Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.075223 4824 scope.go:117] "RemoveContainer" containerID="5ea3d7e5055f26b89248f5118b0972582d7d96bea9dedca1d4da15bbb311e5d7" Dec 09 10:12:40 crc kubenswrapper[4824]: E1209 10:12:40.076119 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ea3d7e5055f26b89248f5118b0972582d7d96bea9dedca1d4da15bbb311e5d7\": container with ID starting with 5ea3d7e5055f26b89248f5118b0972582d7d96bea9dedca1d4da15bbb311e5d7 not found: ID does not exist" containerID="5ea3d7e5055f26b89248f5118b0972582d7d96bea9dedca1d4da15bbb311e5d7" Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.076173 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ea3d7e5055f26b89248f5118b0972582d7d96bea9dedca1d4da15bbb311e5d7"} err="failed to get container status \"5ea3d7e5055f26b89248f5118b0972582d7d96bea9dedca1d4da15bbb311e5d7\": rpc error: code = NotFound desc = could not find container \"5ea3d7e5055f26b89248f5118b0972582d7d96bea9dedca1d4da15bbb311e5d7\": container with ID starting with 5ea3d7e5055f26b89248f5118b0972582d7d96bea9dedca1d4da15bbb311e5d7 not found: ID does not exist" Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.076207 4824 scope.go:117] "RemoveContainer" containerID="b5cf173a26f1977cf61d3b4184fa796463dc2a63ac661f657cca3f3e873af83f" Dec 09 10:12:40 crc kubenswrapper[4824]: E1209 10:12:40.077065 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5cf173a26f1977cf61d3b4184fa796463dc2a63ac661f657cca3f3e873af83f\": container with ID starting with b5cf173a26f1977cf61d3b4184fa796463dc2a63ac661f657cca3f3e873af83f not found: ID does not exist" containerID="b5cf173a26f1977cf61d3b4184fa796463dc2a63ac661f657cca3f3e873af83f" Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.077104 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5cf173a26f1977cf61d3b4184fa796463dc2a63ac661f657cca3f3e873af83f"} err="failed to get container status \"b5cf173a26f1977cf61d3b4184fa796463dc2a63ac661f657cca3f3e873af83f\": rpc error: code = NotFound desc = could not find container \"b5cf173a26f1977cf61d3b4184fa796463dc2a63ac661f657cca3f3e873af83f\": container with ID starting with b5cf173a26f1977cf61d3b4184fa796463dc2a63ac661f657cca3f3e873af83f not found: ID does not exist" Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.131067 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-config\") pod \"26b61717-1239-4782-bc75-6b0eece01c14\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.131306 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prkxz\" (UniqueName: \"kubernetes.io/projected/26b61717-1239-4782-bc75-6b0eece01c14-kube-api-access-prkxz\") pod \"26b61717-1239-4782-bc75-6b0eece01c14\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.131366 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-dns-svc\") pod \"26b61717-1239-4782-bc75-6b0eece01c14\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.131414 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-dns-swift-storage-0\") pod \"26b61717-1239-4782-bc75-6b0eece01c14\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.131500 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-ovsdbserver-sb\") pod \"26b61717-1239-4782-bc75-6b0eece01c14\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.131584 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-ovsdbserver-nb\") pod \"26b61717-1239-4782-bc75-6b0eece01c14\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.146101 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5f9b44b4db-xxc6m"] Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.161398 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26b61717-1239-4782-bc75-6b0eece01c14-kube-api-access-prkxz" (OuterVolumeSpecName: "kube-api-access-prkxz") pod "26b61717-1239-4782-bc75-6b0eece01c14" (UID: "26b61717-1239-4782-bc75-6b0eece01c14"). InnerVolumeSpecName "kube-api-access-prkxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.182917 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5f9b44b4db-xxc6m"] Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.225866 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-config" (OuterVolumeSpecName: "config") pod "26b61717-1239-4782-bc75-6b0eece01c14" (UID: "26b61717-1239-4782-bc75-6b0eece01c14"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.235499 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.235553 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prkxz\" (UniqueName: \"kubernetes.io/projected/26b61717-1239-4782-bc75-6b0eece01c14-kube-api-access-prkxz\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.254674 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "26b61717-1239-4782-bc75-6b0eece01c14" (UID: "26b61717-1239-4782-bc75-6b0eece01c14"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.263210 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "26b61717-1239-4782-bc75-6b0eece01c14" (UID: "26b61717-1239-4782-bc75-6b0eece01c14"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.310363 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "26b61717-1239-4782-bc75-6b0eece01c14" (UID: "26b61717-1239-4782-bc75-6b0eece01c14"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.337255 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "26b61717-1239-4782-bc75-6b0eece01c14" (UID: "26b61717-1239-4782-bc75-6b0eece01c14"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.337504 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-dns-swift-storage-0\") pod \"26b61717-1239-4782-bc75-6b0eece01c14\" (UID: \"26b61717-1239-4782-bc75-6b0eece01c14\") " Dec 09 10:12:40 crc kubenswrapper[4824]: W1209 10:12:40.337754 4824 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/26b61717-1239-4782-bc75-6b0eece01c14/volumes/kubernetes.io~configmap/dns-swift-storage-0 Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.337770 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "26b61717-1239-4782-bc75-6b0eece01c14" (UID: "26b61717-1239-4782-bc75-6b0eece01c14"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.338344 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.338370 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.338381 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.338394 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26b61717-1239-4782-bc75-6b0eece01c14-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.607902 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lhlfr"] Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.608210 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lhlfr" podUID="ad8cfdb4-f7e5-4b30-a1b8-a680b111f788" containerName="registry-server" containerID="cri-o://d2821f5dfa01bb037746d395dcb971a50b3480320fac295f2df8ad81c9fe1e90" gracePeriod=2 Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.880604 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-2952m" Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.880653 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-2952m" event={"ID":"26b61717-1239-4782-bc75-6b0eece01c14","Type":"ContainerDied","Data":"c52022598eb16d1fb834933ce2c44550354a25075891170be311b4944b156302"} Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.882523 4824 scope.go:117] "RemoveContainer" containerID="bc738525e61a76ad921d11262bd0345eb8616de11b3587e7a1de703692188cda" Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.912027 4824 generic.go:334] "Generic (PLEG): container finished" podID="ad8cfdb4-f7e5-4b30-a1b8-a680b111f788" containerID="d2821f5dfa01bb037746d395dcb971a50b3480320fac295f2df8ad81c9fe1e90" exitCode=0 Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.912113 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lhlfr" event={"ID":"ad8cfdb4-f7e5-4b30-a1b8-a680b111f788","Type":"ContainerDied","Data":"d2821f5dfa01bb037746d395dcb971a50b3480320fac295f2df8ad81c9fe1e90"} Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.942921 4824 generic.go:334] "Generic (PLEG): container finished" podID="d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5" containerID="355cf99d0ef56866339c2aca549bfc6cb1fed56cd38dd39873d7fefed6d6e046" exitCode=0 Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.943410 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5","Type":"ContainerDied","Data":"355cf99d0ef56866339c2aca549bfc6cb1fed56cd38dd39873d7fefed6d6e046"} Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.951032 4824 scope.go:117] "RemoveContainer" containerID="ebcec85021517dc9f266f1c774c51ff8974b3d9223f06e0739ebf4773c604928" Dec 09 10:12:40 crc kubenswrapper[4824]: I1209 10:12:40.988069 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-2952m"] Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.023645 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-2952m"] Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.190523 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lhlfr" Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.265537 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nctnx\" (UniqueName: \"kubernetes.io/projected/ad8cfdb4-f7e5-4b30-a1b8-a680b111f788-kube-api-access-nctnx\") pod \"ad8cfdb4-f7e5-4b30-a1b8-a680b111f788\" (UID: \"ad8cfdb4-f7e5-4b30-a1b8-a680b111f788\") " Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.265701 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad8cfdb4-f7e5-4b30-a1b8-a680b111f788-utilities\") pod \"ad8cfdb4-f7e5-4b30-a1b8-a680b111f788\" (UID: \"ad8cfdb4-f7e5-4b30-a1b8-a680b111f788\") " Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.265808 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad8cfdb4-f7e5-4b30-a1b8-a680b111f788-catalog-content\") pod \"ad8cfdb4-f7e5-4b30-a1b8-a680b111f788\" (UID: \"ad8cfdb4-f7e5-4b30-a1b8-a680b111f788\") " Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.267226 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad8cfdb4-f7e5-4b30-a1b8-a680b111f788-utilities" (OuterVolumeSpecName: "utilities") pod "ad8cfdb4-f7e5-4b30-a1b8-a680b111f788" (UID: "ad8cfdb4-f7e5-4b30-a1b8-a680b111f788"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.274934 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad8cfdb4-f7e5-4b30-a1b8-a680b111f788-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.280183 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad8cfdb4-f7e5-4b30-a1b8-a680b111f788-kube-api-access-nctnx" (OuterVolumeSpecName: "kube-api-access-nctnx") pod "ad8cfdb4-f7e5-4b30-a1b8-a680b111f788" (UID: "ad8cfdb4-f7e5-4b30-a1b8-a680b111f788"). InnerVolumeSpecName "kube-api-access-nctnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.353883 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.380900 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nctnx\" (UniqueName: \"kubernetes.io/projected/ad8cfdb4-f7e5-4b30-a1b8-a680b111f788-kube-api-access-nctnx\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.470492 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad8cfdb4-f7e5-4b30-a1b8-a680b111f788-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad8cfdb4-f7e5-4b30-a1b8-a680b111f788" (UID: "ad8cfdb4-f7e5-4b30-a1b8-a680b111f788"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.487794 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad8cfdb4-f7e5-4b30-a1b8-a680b111f788-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.866070 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.892179 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7c7958c67b-wr8zx" Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.903262 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-scripts\") pod \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\" (UID: \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\") " Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.903350 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rpxx\" (UniqueName: \"kubernetes.io/projected/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-kube-api-access-8rpxx\") pod \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\" (UID: \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\") " Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.903540 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-config-data-custom\") pod \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\" (UID: \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\") " Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.903647 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-etc-machine-id\") pod \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\" (UID: \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\") " Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.903826 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-combined-ca-bundle\") pod \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\" (UID: \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\") " Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.903954 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-config-data\") pod \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\" (UID: \"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5\") " Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.904432 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5" (UID: "d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.905313 4824 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.916760 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-kube-api-access-8rpxx" (OuterVolumeSpecName: "kube-api-access-8rpxx") pod "d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5" (UID: "d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5"). InnerVolumeSpecName "kube-api-access-8rpxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.926996 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-scripts" (OuterVolumeSpecName: "scripts") pod "d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5" (UID: "d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.945990 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5" (UID: "d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.992640 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26b61717-1239-4782-bc75-6b0eece01c14" path="/var/lib/kubelet/pods/26b61717-1239-4782-bc75-6b0eece01c14/volumes" Dec 09 10:12:41 crc kubenswrapper[4824]: I1209 10:12:41.994023 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f8a567c-2577-4c66-a75e-e43f7cd66000" path="/var/lib/kubelet/pods/9f8a567c-2577-4c66-a75e-e43f7cd66000/volumes" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.009116 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.009174 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rpxx\" (UniqueName: \"kubernetes.io/projected/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-kube-api-access-8rpxx\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.009188 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.017860 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lhlfr" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.032238 4824 generic.go:334] "Generic (PLEG): container finished" podID="d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5" containerID="9a7e126bfed9f9eccf9686b200bfbceca110dcae02d5023b921a7b86b96e0d92" exitCode=0 Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.032550 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.067216 4824 generic.go:334] "Generic (PLEG): container finished" podID="065686b2-246c-4194-936f-b6e58ee9763b" containerID="8d9fe5540c0ecc67d4aa35f991a15750e270b158cfe77e9cdbdfaebd836fa3c7" exitCode=0 Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.149535 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lhlfr" event={"ID":"ad8cfdb4-f7e5-4b30-a1b8-a680b111f788","Type":"ContainerDied","Data":"68a95d250d8635847290cd9ee6d6506d855bd88c48f7ef78ab67805ad6aa0cb1"} Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.149582 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5","Type":"ContainerDied","Data":"9a7e126bfed9f9eccf9686b200bfbceca110dcae02d5023b921a7b86b96e0d92"} Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.149603 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5","Type":"ContainerDied","Data":"a81cdeb105fa88f95542073b2f7f749454390584714885ceb38ec6192c185f46"} Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.149613 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wx2h9" event={"ID":"065686b2-246c-4194-936f-b6e58ee9763b","Type":"ContainerDied","Data":"8d9fe5540c0ecc67d4aa35f991a15750e270b158cfe77e9cdbdfaebd836fa3c7"} Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.149638 4824 scope.go:117] "RemoveContainer" containerID="d2821f5dfa01bb037746d395dcb971a50b3480320fac295f2df8ad81c9fe1e90" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.179761 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5" (UID: "d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.211313 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-config-data" (OuterVolumeSpecName: "config-data") pod "d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5" (UID: "d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.215166 4824 scope.go:117] "RemoveContainer" containerID="c0650433deec5abea39f2881e472191bd159aadb67a9c737ba5bd6be7c0a63c9" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.222578 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lhlfr"] Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.223909 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.223944 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.238161 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lhlfr"] Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.264724 4824 scope.go:117] "RemoveContainer" containerID="cc5def6a5a7d96bf7694ed1229224b7871dd16799c2c595a8b42dad978e0ca13" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.301128 4824 scope.go:117] "RemoveContainer" containerID="355cf99d0ef56866339c2aca549bfc6cb1fed56cd38dd39873d7fefed6d6e046" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.331882 4824 scope.go:117] "RemoveContainer" containerID="9a7e126bfed9f9eccf9686b200bfbceca110dcae02d5023b921a7b86b96e0d92" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.363743 4824 scope.go:117] "RemoveContainer" containerID="355cf99d0ef56866339c2aca549bfc6cb1fed56cd38dd39873d7fefed6d6e046" Dec 09 10:12:42 crc kubenswrapper[4824]: E1209 10:12:42.364562 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"355cf99d0ef56866339c2aca549bfc6cb1fed56cd38dd39873d7fefed6d6e046\": container with ID starting with 355cf99d0ef56866339c2aca549bfc6cb1fed56cd38dd39873d7fefed6d6e046 not found: ID does not exist" containerID="355cf99d0ef56866339c2aca549bfc6cb1fed56cd38dd39873d7fefed6d6e046" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.364606 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"355cf99d0ef56866339c2aca549bfc6cb1fed56cd38dd39873d7fefed6d6e046"} err="failed to get container status \"355cf99d0ef56866339c2aca549bfc6cb1fed56cd38dd39873d7fefed6d6e046\": rpc error: code = NotFound desc = could not find container \"355cf99d0ef56866339c2aca549bfc6cb1fed56cd38dd39873d7fefed6d6e046\": container with ID starting with 355cf99d0ef56866339c2aca549bfc6cb1fed56cd38dd39873d7fefed6d6e046 not found: ID does not exist" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.364647 4824 scope.go:117] "RemoveContainer" containerID="9a7e126bfed9f9eccf9686b200bfbceca110dcae02d5023b921a7b86b96e0d92" Dec 09 10:12:42 crc kubenswrapper[4824]: E1209 10:12:42.364991 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a7e126bfed9f9eccf9686b200bfbceca110dcae02d5023b921a7b86b96e0d92\": container with ID starting with 9a7e126bfed9f9eccf9686b200bfbceca110dcae02d5023b921a7b86b96e0d92 not found: ID does not exist" containerID="9a7e126bfed9f9eccf9686b200bfbceca110dcae02d5023b921a7b86b96e0d92" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.365045 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a7e126bfed9f9eccf9686b200bfbceca110dcae02d5023b921a7b86b96e0d92"} err="failed to get container status \"9a7e126bfed9f9eccf9686b200bfbceca110dcae02d5023b921a7b86b96e0d92\": rpc error: code = NotFound desc = could not find container \"9a7e126bfed9f9eccf9686b200bfbceca110dcae02d5023b921a7b86b96e0d92\": container with ID starting with 9a7e126bfed9f9eccf9686b200bfbceca110dcae02d5023b921a7b86b96e0d92 not found: ID does not exist" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.395940 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.414536 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.434403 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 10:12:42 crc kubenswrapper[4824]: E1209 10:12:42.435141 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad8cfdb4-f7e5-4b30-a1b8-a680b111f788" containerName="extract-utilities" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.435540 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad8cfdb4-f7e5-4b30-a1b8-a680b111f788" containerName="extract-utilities" Dec 09 10:12:42 crc kubenswrapper[4824]: E1209 10:12:42.435564 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f8a567c-2577-4c66-a75e-e43f7cd66000" containerName="barbican-api" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.435573 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f8a567c-2577-4c66-a75e-e43f7cd66000" containerName="barbican-api" Dec 09 10:12:42 crc kubenswrapper[4824]: E1209 10:12:42.435593 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad8cfdb4-f7e5-4b30-a1b8-a680b111f788" containerName="extract-content" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.435601 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad8cfdb4-f7e5-4b30-a1b8-a680b111f788" containerName="extract-content" Dec 09 10:12:42 crc kubenswrapper[4824]: E1209 10:12:42.435625 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5" containerName="probe" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.435632 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5" containerName="probe" Dec 09 10:12:42 crc kubenswrapper[4824]: E1209 10:12:42.435652 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b61717-1239-4782-bc75-6b0eece01c14" containerName="init" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.435664 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b61717-1239-4782-bc75-6b0eece01c14" containerName="init" Dec 09 10:12:42 crc kubenswrapper[4824]: E1209 10:12:42.435681 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f8a567c-2577-4c66-a75e-e43f7cd66000" containerName="barbican-api-log" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.435690 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f8a567c-2577-4c66-a75e-e43f7cd66000" containerName="barbican-api-log" Dec 09 10:12:42 crc kubenswrapper[4824]: E1209 10:12:42.435729 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5" containerName="cinder-scheduler" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.435738 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5" containerName="cinder-scheduler" Dec 09 10:12:42 crc kubenswrapper[4824]: E1209 10:12:42.435754 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad8cfdb4-f7e5-4b30-a1b8-a680b111f788" containerName="registry-server" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.435763 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad8cfdb4-f7e5-4b30-a1b8-a680b111f788" containerName="registry-server" Dec 09 10:12:42 crc kubenswrapper[4824]: E1209 10:12:42.435810 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b61717-1239-4782-bc75-6b0eece01c14" containerName="dnsmasq-dns" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.435818 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b61717-1239-4782-bc75-6b0eece01c14" containerName="dnsmasq-dns" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.436140 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5" containerName="probe" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.436162 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f8a567c-2577-4c66-a75e-e43f7cd66000" containerName="barbican-api-log" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.436183 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b61717-1239-4782-bc75-6b0eece01c14" containerName="dnsmasq-dns" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.436203 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f8a567c-2577-4c66-a75e-e43f7cd66000" containerName="barbican-api" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.436230 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad8cfdb4-f7e5-4b30-a1b8-a680b111f788" containerName="registry-server" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.436254 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5" containerName="cinder-scheduler" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.481716 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.493647 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.497548 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.529843 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/153fe73c-2767-4f80-a7b4-0af49257065a-config-data\") pod \"cinder-scheduler-0\" (UID: \"153fe73c-2767-4f80-a7b4-0af49257065a\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.529902 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/153fe73c-2767-4f80-a7b4-0af49257065a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"153fe73c-2767-4f80-a7b4-0af49257065a\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.529987 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/153fe73c-2767-4f80-a7b4-0af49257065a-scripts\") pod \"cinder-scheduler-0\" (UID: \"153fe73c-2767-4f80-a7b4-0af49257065a\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.530056 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/153fe73c-2767-4f80-a7b4-0af49257065a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"153fe73c-2767-4f80-a7b4-0af49257065a\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.530268 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/153fe73c-2767-4f80-a7b4-0af49257065a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"153fe73c-2767-4f80-a7b4-0af49257065a\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.530409 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7kvq\" (UniqueName: \"kubernetes.io/projected/153fe73c-2767-4f80-a7b4-0af49257065a-kube-api-access-f7kvq\") pod \"cinder-scheduler-0\" (UID: \"153fe73c-2767-4f80-a7b4-0af49257065a\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.636227 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/153fe73c-2767-4f80-a7b4-0af49257065a-scripts\") pod \"cinder-scheduler-0\" (UID: \"153fe73c-2767-4f80-a7b4-0af49257065a\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.637214 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/153fe73c-2767-4f80-a7b4-0af49257065a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"153fe73c-2767-4f80-a7b4-0af49257065a\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.637318 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/153fe73c-2767-4f80-a7b4-0af49257065a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"153fe73c-2767-4f80-a7b4-0af49257065a\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.637419 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7kvq\" (UniqueName: \"kubernetes.io/projected/153fe73c-2767-4f80-a7b4-0af49257065a-kube-api-access-f7kvq\") pod \"cinder-scheduler-0\" (UID: \"153fe73c-2767-4f80-a7b4-0af49257065a\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.637472 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/153fe73c-2767-4f80-a7b4-0af49257065a-config-data\") pod \"cinder-scheduler-0\" (UID: \"153fe73c-2767-4f80-a7b4-0af49257065a\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.637503 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/153fe73c-2767-4f80-a7b4-0af49257065a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"153fe73c-2767-4f80-a7b4-0af49257065a\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.637684 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/153fe73c-2767-4f80-a7b4-0af49257065a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"153fe73c-2767-4f80-a7b4-0af49257065a\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.642268 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/153fe73c-2767-4f80-a7b4-0af49257065a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"153fe73c-2767-4f80-a7b4-0af49257065a\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.642283 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/153fe73c-2767-4f80-a7b4-0af49257065a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"153fe73c-2767-4f80-a7b4-0af49257065a\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.642357 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/153fe73c-2767-4f80-a7b4-0af49257065a-scripts\") pod \"cinder-scheduler-0\" (UID: \"153fe73c-2767-4f80-a7b4-0af49257065a\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.647101 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/153fe73c-2767-4f80-a7b4-0af49257065a-config-data\") pod \"cinder-scheduler-0\" (UID: \"153fe73c-2767-4f80-a7b4-0af49257065a\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.662594 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7kvq\" (UniqueName: \"kubernetes.io/projected/153fe73c-2767-4f80-a7b4-0af49257065a-kube-api-access-f7kvq\") pod \"cinder-scheduler-0\" (UID: \"153fe73c-2767-4f80-a7b4-0af49257065a\") " pod="openstack/cinder-scheduler-0" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.825493 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-9c7879457-2fmzw" Dec 09 10:12:42 crc kubenswrapper[4824]: I1209 10:12:42.895927 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 10:12:43 crc kubenswrapper[4824]: I1209 10:12:43.093062 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wx2h9" event={"ID":"065686b2-246c-4194-936f-b6e58ee9763b","Type":"ContainerStarted","Data":"2d424024ddebda113c5fbcb03177ec44cb898168337f7d67047fc5844fbd2165"} Dec 09 10:12:43 crc kubenswrapper[4824]: I1209 10:12:43.125958 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wx2h9" podStartSLOduration=3.44435213 podStartE2EDuration="6.125933333s" podCreationTimestamp="2025-12-09 10:12:37 +0000 UTC" firstStartedPulling="2025-12-09 10:12:39.872007343 +0000 UTC m=+1516.206512010" lastFinishedPulling="2025-12-09 10:12:42.553588546 +0000 UTC m=+1518.888093213" observedRunningTime="2025-12-09 10:12:43.115071208 +0000 UTC m=+1519.449575895" watchObservedRunningTime="2025-12-09 10:12:43.125933333 +0000 UTC m=+1519.460438000" Dec 09 10:12:43 crc kubenswrapper[4824]: I1209 10:12:43.134078 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 09 10:12:43 crc kubenswrapper[4824]: W1209 10:12:43.518983 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod153fe73c_2767_4f80_a7b4_0af49257065a.slice/crio-200610107b035f6cb62c7749b49841543d9b3c527e9e9b373aa2ad46cff83db7 WatchSource:0}: Error finding container 200610107b035f6cb62c7749b49841543d9b3c527e9e9b373aa2ad46cff83db7: Status 404 returned error can't find the container with id 200610107b035f6cb62c7749b49841543d9b3c527e9e9b373aa2ad46cff83db7 Dec 09 10:12:43 crc kubenswrapper[4824]: I1209 10:12:43.522684 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 10:12:43 crc kubenswrapper[4824]: I1209 10:12:43.929732 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad8cfdb4-f7e5-4b30-a1b8-a680b111f788" path="/var/lib/kubelet/pods/ad8cfdb4-f7e5-4b30-a1b8-a680b111f788/volumes" Dec 09 10:12:43 crc kubenswrapper[4824]: I1209 10:12:43.931222 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5" path="/var/lib/kubelet/pods/d87ffb93-4ef1-4f9d-973f-0c637c7cb8f5/volumes" Dec 09 10:12:44 crc kubenswrapper[4824]: I1209 10:12:44.155262 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"153fe73c-2767-4f80-a7b4-0af49257065a","Type":"ContainerStarted","Data":"200610107b035f6cb62c7749b49841543d9b3c527e9e9b373aa2ad46cff83db7"} Dec 09 10:12:44 crc kubenswrapper[4824]: I1209 10:12:44.580944 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-85ff748b95-2952m" podUID="26b61717-1239-4782-bc75-6b0eece01c14" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.200:5353: i/o timeout" Dec 09 10:12:45 crc kubenswrapper[4824]: I1209 10:12:45.168496 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"153fe73c-2767-4f80-a7b4-0af49257065a","Type":"ContainerStarted","Data":"d2f14825848e7e355909e904cb07d3ac37df2eec491e11d6f97757df6a0f4a7d"} Dec 09 10:12:45 crc kubenswrapper[4824]: I1209 10:12:45.959485 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 09 10:12:45 crc kubenswrapper[4824]: I1209 10:12:45.965443 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 09 10:12:45 crc kubenswrapper[4824]: I1209 10:12:45.979218 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-26k7j" Dec 09 10:12:45 crc kubenswrapper[4824]: I1209 10:12:45.979563 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 09 10:12:45 crc kubenswrapper[4824]: I1209 10:12:45.979826 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 09 10:12:46 crc kubenswrapper[4824]: I1209 10:12:46.011923 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 09 10:12:46 crc kubenswrapper[4824]: I1209 10:12:46.154978 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a00596c6-582b-4ec7-9920-6dddc8197001-combined-ca-bundle\") pod \"openstackclient\" (UID: \"a00596c6-582b-4ec7-9920-6dddc8197001\") " pod="openstack/openstackclient" Dec 09 10:12:46 crc kubenswrapper[4824]: I1209 10:12:46.155223 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a00596c6-582b-4ec7-9920-6dddc8197001-openstack-config-secret\") pod \"openstackclient\" (UID: \"a00596c6-582b-4ec7-9920-6dddc8197001\") " pod="openstack/openstackclient" Dec 09 10:12:46 crc kubenswrapper[4824]: I1209 10:12:46.155744 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a00596c6-582b-4ec7-9920-6dddc8197001-openstack-config\") pod \"openstackclient\" (UID: \"a00596c6-582b-4ec7-9920-6dddc8197001\") " pod="openstack/openstackclient" Dec 09 10:12:46 crc kubenswrapper[4824]: I1209 10:12:46.155985 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scxsc\" (UniqueName: \"kubernetes.io/projected/a00596c6-582b-4ec7-9920-6dddc8197001-kube-api-access-scxsc\") pod \"openstackclient\" (UID: \"a00596c6-582b-4ec7-9920-6dddc8197001\") " pod="openstack/openstackclient" Dec 09 10:12:46 crc kubenswrapper[4824]: I1209 10:12:46.182534 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"153fe73c-2767-4f80-a7b4-0af49257065a","Type":"ContainerStarted","Data":"8818c1aa97e1dd10241a60a57ab1fe8eec210334503b76a34dd05dc1d093b55b"} Dec 09 10:12:46 crc kubenswrapper[4824]: I1209 10:12:46.211799 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.211754384 podStartE2EDuration="4.211754384s" podCreationTimestamp="2025-12-09 10:12:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:12:46.203139419 +0000 UTC m=+1522.537644106" watchObservedRunningTime="2025-12-09 10:12:46.211754384 +0000 UTC m=+1522.546259051" Dec 09 10:12:46 crc kubenswrapper[4824]: I1209 10:12:46.258158 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a00596c6-582b-4ec7-9920-6dddc8197001-combined-ca-bundle\") pod \"openstackclient\" (UID: \"a00596c6-582b-4ec7-9920-6dddc8197001\") " pod="openstack/openstackclient" Dec 09 10:12:46 crc kubenswrapper[4824]: I1209 10:12:46.258251 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a00596c6-582b-4ec7-9920-6dddc8197001-openstack-config-secret\") pod \"openstackclient\" (UID: \"a00596c6-582b-4ec7-9920-6dddc8197001\") " pod="openstack/openstackclient" Dec 09 10:12:46 crc kubenswrapper[4824]: I1209 10:12:46.258411 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a00596c6-582b-4ec7-9920-6dddc8197001-openstack-config\") pod \"openstackclient\" (UID: \"a00596c6-582b-4ec7-9920-6dddc8197001\") " pod="openstack/openstackclient" Dec 09 10:12:46 crc kubenswrapper[4824]: I1209 10:12:46.258481 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scxsc\" (UniqueName: \"kubernetes.io/projected/a00596c6-582b-4ec7-9920-6dddc8197001-kube-api-access-scxsc\") pod \"openstackclient\" (UID: \"a00596c6-582b-4ec7-9920-6dddc8197001\") " pod="openstack/openstackclient" Dec 09 10:12:46 crc kubenswrapper[4824]: I1209 10:12:46.259703 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a00596c6-582b-4ec7-9920-6dddc8197001-openstack-config\") pod \"openstackclient\" (UID: \"a00596c6-582b-4ec7-9920-6dddc8197001\") " pod="openstack/openstackclient" Dec 09 10:12:46 crc kubenswrapper[4824]: I1209 10:12:46.265486 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a00596c6-582b-4ec7-9920-6dddc8197001-openstack-config-secret\") pod \"openstackclient\" (UID: \"a00596c6-582b-4ec7-9920-6dddc8197001\") " pod="openstack/openstackclient" Dec 09 10:12:46 crc kubenswrapper[4824]: I1209 10:12:46.275673 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a00596c6-582b-4ec7-9920-6dddc8197001-combined-ca-bundle\") pod \"openstackclient\" (UID: \"a00596c6-582b-4ec7-9920-6dddc8197001\") " pod="openstack/openstackclient" Dec 09 10:12:46 crc kubenswrapper[4824]: I1209 10:12:46.278348 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scxsc\" (UniqueName: \"kubernetes.io/projected/a00596c6-582b-4ec7-9920-6dddc8197001-kube-api-access-scxsc\") pod \"openstackclient\" (UID: \"a00596c6-582b-4ec7-9920-6dddc8197001\") " pod="openstack/openstackclient" Dec 09 10:12:46 crc kubenswrapper[4824]: I1209 10:12:46.322147 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 09 10:12:46 crc kubenswrapper[4824]: I1209 10:12:46.969564 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 09 10:12:47 crc kubenswrapper[4824]: I1209 10:12:47.196841 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"a00596c6-582b-4ec7-9920-6dddc8197001","Type":"ContainerStarted","Data":"82593b8f2a1c5afc79bd4bbca51c59da249b5f26845cec90867f59a61ef4c4d1"} Dec 09 10:12:47 crc kubenswrapper[4824]: I1209 10:12:47.896900 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 09 10:12:48 crc kubenswrapper[4824]: I1209 10:12:48.035512 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wx2h9" Dec 09 10:12:48 crc kubenswrapper[4824]: I1209 10:12:48.035574 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wx2h9" Dec 09 10:12:48 crc kubenswrapper[4824]: I1209 10:12:48.114921 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wx2h9" Dec 09 10:12:48 crc kubenswrapper[4824]: I1209 10:12:48.293492 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wx2h9" Dec 09 10:12:48 crc kubenswrapper[4824]: I1209 10:12:48.372149 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wx2h9"] Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.253796 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wx2h9" podUID="065686b2-246c-4194-936f-b6e58ee9763b" containerName="registry-server" containerID="cri-o://2d424024ddebda113c5fbcb03177ec44cb898168337f7d67047fc5844fbd2165" gracePeriod=2 Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.654634 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-6fffc89c95-wt6ms"] Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.672032 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6fffc89c95-wt6ms" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.681634 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-k8rrz" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.683606 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.687152 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.688697 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04bcf8d0-9439-4739-a503-e504651a80dd-config-data\") pod \"heat-engine-6fffc89c95-wt6ms\" (UID: \"04bcf8d0-9439-4739-a503-e504651a80dd\") " pod="openstack/heat-engine-6fffc89c95-wt6ms" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.688735 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04bcf8d0-9439-4739-a503-e504651a80dd-config-data-custom\") pod \"heat-engine-6fffc89c95-wt6ms\" (UID: \"04bcf8d0-9439-4739-a503-e504651a80dd\") " pod="openstack/heat-engine-6fffc89c95-wt6ms" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.688767 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27qpm\" (UniqueName: \"kubernetes.io/projected/04bcf8d0-9439-4739-a503-e504651a80dd-kube-api-access-27qpm\") pod \"heat-engine-6fffc89c95-wt6ms\" (UID: \"04bcf8d0-9439-4739-a503-e504651a80dd\") " pod="openstack/heat-engine-6fffc89c95-wt6ms" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.688809 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04bcf8d0-9439-4739-a503-e504651a80dd-combined-ca-bundle\") pod \"heat-engine-6fffc89c95-wt6ms\" (UID: \"04bcf8d0-9439-4739-a503-e504651a80dd\") " pod="openstack/heat-engine-6fffc89c95-wt6ms" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.715038 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6fffc89c95-wt6ms"] Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.793048 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04bcf8d0-9439-4739-a503-e504651a80dd-config-data\") pod \"heat-engine-6fffc89c95-wt6ms\" (UID: \"04bcf8d0-9439-4739-a503-e504651a80dd\") " pod="openstack/heat-engine-6fffc89c95-wt6ms" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.793124 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04bcf8d0-9439-4739-a503-e504651a80dd-config-data-custom\") pod \"heat-engine-6fffc89c95-wt6ms\" (UID: \"04bcf8d0-9439-4739-a503-e504651a80dd\") " pod="openstack/heat-engine-6fffc89c95-wt6ms" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.793152 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27qpm\" (UniqueName: \"kubernetes.io/projected/04bcf8d0-9439-4739-a503-e504651a80dd-kube-api-access-27qpm\") pod \"heat-engine-6fffc89c95-wt6ms\" (UID: \"04bcf8d0-9439-4739-a503-e504651a80dd\") " pod="openstack/heat-engine-6fffc89c95-wt6ms" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.793186 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04bcf8d0-9439-4739-a503-e504651a80dd-combined-ca-bundle\") pod \"heat-engine-6fffc89c95-wt6ms\" (UID: \"04bcf8d0-9439-4739-a503-e504651a80dd\") " pod="openstack/heat-engine-6fffc89c95-wt6ms" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.805032 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-tcz7b"] Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.807385 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.818170 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-7466b579b4-qpdj6"] Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.820008 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7466b579b4-qpdj6" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.821792 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04bcf8d0-9439-4739-a503-e504651a80dd-config-data-custom\") pod \"heat-engine-6fffc89c95-wt6ms\" (UID: \"04bcf8d0-9439-4739-a503-e504651a80dd\") " pod="openstack/heat-engine-6fffc89c95-wt6ms" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.825036 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04bcf8d0-9439-4739-a503-e504651a80dd-config-data\") pod \"heat-engine-6fffc89c95-wt6ms\" (UID: \"04bcf8d0-9439-4739-a503-e504651a80dd\") " pod="openstack/heat-engine-6fffc89c95-wt6ms" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.825805 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.846071 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04bcf8d0-9439-4739-a503-e504651a80dd-combined-ca-bundle\") pod \"heat-engine-6fffc89c95-wt6ms\" (UID: \"04bcf8d0-9439-4739-a503-e504651a80dd\") " pod="openstack/heat-engine-6fffc89c95-wt6ms" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.849908 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-tcz7b"] Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.873616 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27qpm\" (UniqueName: \"kubernetes.io/projected/04bcf8d0-9439-4739-a503-e504651a80dd-kube-api-access-27qpm\") pod \"heat-engine-6fffc89c95-wt6ms\" (UID: \"04bcf8d0-9439-4739-a503-e504651a80dd\") " pod="openstack/heat-engine-6fffc89c95-wt6ms" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.902583 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e74eb53-a9de-4679-99d1-7bb097c01d93-combined-ca-bundle\") pod \"heat-cfnapi-7466b579b4-qpdj6\" (UID: \"0e74eb53-a9de-4679-99d1-7bb097c01d93\") " pod="openstack/heat-cfnapi-7466b579b4-qpdj6" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.902644 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-tcz7b\" (UID: \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\") " pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.902714 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-tcz7b\" (UID: \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\") " pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.902745 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l79k7\" (UniqueName: \"kubernetes.io/projected/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-kube-api-access-l79k7\") pod \"dnsmasq-dns-7756b9d78c-tcz7b\" (UID: \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\") " pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.902807 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-config\") pod \"dnsmasq-dns-7756b9d78c-tcz7b\" (UID: \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\") " pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.902847 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e74eb53-a9de-4679-99d1-7bb097c01d93-config-data\") pod \"heat-cfnapi-7466b579b4-qpdj6\" (UID: \"0e74eb53-a9de-4679-99d1-7bb097c01d93\") " pod="openstack/heat-cfnapi-7466b579b4-qpdj6" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.902864 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlg58\" (UniqueName: \"kubernetes.io/projected/0e74eb53-a9de-4679-99d1-7bb097c01d93-kube-api-access-tlg58\") pod \"heat-cfnapi-7466b579b4-qpdj6\" (UID: \"0e74eb53-a9de-4679-99d1-7bb097c01d93\") " pod="openstack/heat-cfnapi-7466b579b4-qpdj6" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.902936 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e74eb53-a9de-4679-99d1-7bb097c01d93-config-data-custom\") pod \"heat-cfnapi-7466b579b4-qpdj6\" (UID: \"0e74eb53-a9de-4679-99d1-7bb097c01d93\") " pod="openstack/heat-cfnapi-7466b579b4-qpdj6" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.902990 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-tcz7b\" (UID: \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\") " pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.903039 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-tcz7b\" (UID: \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\") " pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" Dec 09 10:12:50 crc kubenswrapper[4824]: I1209 10:12:50.956876 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7466b579b4-qpdj6"] Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.006427 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6fffc89c95-wt6ms" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.008344 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e74eb53-a9de-4679-99d1-7bb097c01d93-combined-ca-bundle\") pod \"heat-cfnapi-7466b579b4-qpdj6\" (UID: \"0e74eb53-a9de-4679-99d1-7bb097c01d93\") " pod="openstack/heat-cfnapi-7466b579b4-qpdj6" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.008381 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-tcz7b\" (UID: \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\") " pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.008454 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-tcz7b\" (UID: \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\") " pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.008475 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l79k7\" (UniqueName: \"kubernetes.io/projected/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-kube-api-access-l79k7\") pod \"dnsmasq-dns-7756b9d78c-tcz7b\" (UID: \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\") " pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.008530 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-config\") pod \"dnsmasq-dns-7756b9d78c-tcz7b\" (UID: \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\") " pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.008570 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e74eb53-a9de-4679-99d1-7bb097c01d93-config-data\") pod \"heat-cfnapi-7466b579b4-qpdj6\" (UID: \"0e74eb53-a9de-4679-99d1-7bb097c01d93\") " pod="openstack/heat-cfnapi-7466b579b4-qpdj6" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.008589 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlg58\" (UniqueName: \"kubernetes.io/projected/0e74eb53-a9de-4679-99d1-7bb097c01d93-kube-api-access-tlg58\") pod \"heat-cfnapi-7466b579b4-qpdj6\" (UID: \"0e74eb53-a9de-4679-99d1-7bb097c01d93\") " pod="openstack/heat-cfnapi-7466b579b4-qpdj6" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.008709 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e74eb53-a9de-4679-99d1-7bb097c01d93-config-data-custom\") pod \"heat-cfnapi-7466b579b4-qpdj6\" (UID: \"0e74eb53-a9de-4679-99d1-7bb097c01d93\") " pod="openstack/heat-cfnapi-7466b579b4-qpdj6" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.008753 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-tcz7b\" (UID: \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\") " pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.008807 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-tcz7b\" (UID: \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\") " pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.023111 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-tcz7b\" (UID: \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\") " pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.023358 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e74eb53-a9de-4679-99d1-7bb097c01d93-combined-ca-bundle\") pod \"heat-cfnapi-7466b579b4-qpdj6\" (UID: \"0e74eb53-a9de-4679-99d1-7bb097c01d93\") " pod="openstack/heat-cfnapi-7466b579b4-qpdj6" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.024111 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-config\") pod \"dnsmasq-dns-7756b9d78c-tcz7b\" (UID: \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\") " pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.024397 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-tcz7b\" (UID: \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\") " pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.024506 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-tcz7b\" (UID: \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\") " pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.024705 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-tcz7b\" (UID: \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\") " pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.051140 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e74eb53-a9de-4679-99d1-7bb097c01d93-config-data\") pod \"heat-cfnapi-7466b579b4-qpdj6\" (UID: \"0e74eb53-a9de-4679-99d1-7bb097c01d93\") " pod="openstack/heat-cfnapi-7466b579b4-qpdj6" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.051861 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e74eb53-a9de-4679-99d1-7bb097c01d93-config-data-custom\") pod \"heat-cfnapi-7466b579b4-qpdj6\" (UID: \"0e74eb53-a9de-4679-99d1-7bb097c01d93\") " pod="openstack/heat-cfnapi-7466b579b4-qpdj6" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.055775 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l79k7\" (UniqueName: \"kubernetes.io/projected/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-kube-api-access-l79k7\") pod \"dnsmasq-dns-7756b9d78c-tcz7b\" (UID: \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\") " pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.058797 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-79457998d5-fpwv7"] Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.061030 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-79457998d5-fpwv7" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.067246 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlg58\" (UniqueName: \"kubernetes.io/projected/0e74eb53-a9de-4679-99d1-7bb097c01d93-kube-api-access-tlg58\") pod \"heat-cfnapi-7466b579b4-qpdj6\" (UID: \"0e74eb53-a9de-4679-99d1-7bb097c01d93\") " pod="openstack/heat-cfnapi-7466b579b4-qpdj6" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.068299 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.074218 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-79457998d5-fpwv7"] Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.113205 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdlzz\" (UniqueName: \"kubernetes.io/projected/aeb18289-ca5e-45d5-8976-53d77b8c8ee9-kube-api-access-vdlzz\") pod \"heat-api-79457998d5-fpwv7\" (UID: \"aeb18289-ca5e-45d5-8976-53d77b8c8ee9\") " pod="openstack/heat-api-79457998d5-fpwv7" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.113330 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aeb18289-ca5e-45d5-8976-53d77b8c8ee9-config-data\") pod \"heat-api-79457998d5-fpwv7\" (UID: \"aeb18289-ca5e-45d5-8976-53d77b8c8ee9\") " pod="openstack/heat-api-79457998d5-fpwv7" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.113485 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aeb18289-ca5e-45d5-8976-53d77b8c8ee9-config-data-custom\") pod \"heat-api-79457998d5-fpwv7\" (UID: \"aeb18289-ca5e-45d5-8976-53d77b8c8ee9\") " pod="openstack/heat-api-79457998d5-fpwv7" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.113548 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aeb18289-ca5e-45d5-8976-53d77b8c8ee9-combined-ca-bundle\") pod \"heat-api-79457998d5-fpwv7\" (UID: \"aeb18289-ca5e-45d5-8976-53d77b8c8ee9\") " pod="openstack/heat-api-79457998d5-fpwv7" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.216999 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aeb18289-ca5e-45d5-8976-53d77b8c8ee9-config-data-custom\") pod \"heat-api-79457998d5-fpwv7\" (UID: \"aeb18289-ca5e-45d5-8976-53d77b8c8ee9\") " pod="openstack/heat-api-79457998d5-fpwv7" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.217098 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aeb18289-ca5e-45d5-8976-53d77b8c8ee9-combined-ca-bundle\") pod \"heat-api-79457998d5-fpwv7\" (UID: \"aeb18289-ca5e-45d5-8976-53d77b8c8ee9\") " pod="openstack/heat-api-79457998d5-fpwv7" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.217717 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdlzz\" (UniqueName: \"kubernetes.io/projected/aeb18289-ca5e-45d5-8976-53d77b8c8ee9-kube-api-access-vdlzz\") pod \"heat-api-79457998d5-fpwv7\" (UID: \"aeb18289-ca5e-45d5-8976-53d77b8c8ee9\") " pod="openstack/heat-api-79457998d5-fpwv7" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.217912 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aeb18289-ca5e-45d5-8976-53d77b8c8ee9-config-data\") pod \"heat-api-79457998d5-fpwv7\" (UID: \"aeb18289-ca5e-45d5-8976-53d77b8c8ee9\") " pod="openstack/heat-api-79457998d5-fpwv7" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.225596 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aeb18289-ca5e-45d5-8976-53d77b8c8ee9-combined-ca-bundle\") pod \"heat-api-79457998d5-fpwv7\" (UID: \"aeb18289-ca5e-45d5-8976-53d77b8c8ee9\") " pod="openstack/heat-api-79457998d5-fpwv7" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.225994 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aeb18289-ca5e-45d5-8976-53d77b8c8ee9-config-data\") pod \"heat-api-79457998d5-fpwv7\" (UID: \"aeb18289-ca5e-45d5-8976-53d77b8c8ee9\") " pod="openstack/heat-api-79457998d5-fpwv7" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.226323 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aeb18289-ca5e-45d5-8976-53d77b8c8ee9-config-data-custom\") pod \"heat-api-79457998d5-fpwv7\" (UID: \"aeb18289-ca5e-45d5-8976-53d77b8c8ee9\") " pod="openstack/heat-api-79457998d5-fpwv7" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.252639 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdlzz\" (UniqueName: \"kubernetes.io/projected/aeb18289-ca5e-45d5-8976-53d77b8c8ee9-kube-api-access-vdlzz\") pod \"heat-api-79457998d5-fpwv7\" (UID: \"aeb18289-ca5e-45d5-8976-53d77b8c8ee9\") " pod="openstack/heat-api-79457998d5-fpwv7" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.287300 4824 generic.go:334] "Generic (PLEG): container finished" podID="065686b2-246c-4194-936f-b6e58ee9763b" containerID="2d424024ddebda113c5fbcb03177ec44cb898168337f7d67047fc5844fbd2165" exitCode=0 Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.287448 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wx2h9" event={"ID":"065686b2-246c-4194-936f-b6e58ee9763b","Type":"ContainerDied","Data":"2d424024ddebda113c5fbcb03177ec44cb898168337f7d67047fc5844fbd2165"} Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.301441 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.323302 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7466b579b4-qpdj6" Dec 09 10:12:51 crc kubenswrapper[4824]: I1209 10:12:51.466959 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-79457998d5-fpwv7" Dec 09 10:12:52 crc kubenswrapper[4824]: I1209 10:12:52.051336 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6fffc89c95-wt6ms"] Dec 09 10:12:52 crc kubenswrapper[4824]: I1209 10:12:52.051878 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wx2h9" Dec 09 10:12:52 crc kubenswrapper[4824]: W1209 10:12:52.111204 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04bcf8d0_9439_4739_a503_e504651a80dd.slice/crio-ad73bbffb835406c8227acbaa03f03e4ddfc98ccf531b5e7a4e00a36acc7794a WatchSource:0}: Error finding container ad73bbffb835406c8227acbaa03f03e4ddfc98ccf531b5e7a4e00a36acc7794a: Status 404 returned error can't find the container with id ad73bbffb835406c8227acbaa03f03e4ddfc98ccf531b5e7a4e00a36acc7794a Dec 09 10:12:52 crc kubenswrapper[4824]: I1209 10:12:52.128666 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/065686b2-246c-4194-936f-b6e58ee9763b-utilities\") pod \"065686b2-246c-4194-936f-b6e58ee9763b\" (UID: \"065686b2-246c-4194-936f-b6e58ee9763b\") " Dec 09 10:12:52 crc kubenswrapper[4824]: I1209 10:12:52.128863 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwfmq\" (UniqueName: \"kubernetes.io/projected/065686b2-246c-4194-936f-b6e58ee9763b-kube-api-access-bwfmq\") pod \"065686b2-246c-4194-936f-b6e58ee9763b\" (UID: \"065686b2-246c-4194-936f-b6e58ee9763b\") " Dec 09 10:12:52 crc kubenswrapper[4824]: I1209 10:12:52.129060 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/065686b2-246c-4194-936f-b6e58ee9763b-catalog-content\") pod \"065686b2-246c-4194-936f-b6e58ee9763b\" (UID: \"065686b2-246c-4194-936f-b6e58ee9763b\") " Dec 09 10:12:52 crc kubenswrapper[4824]: I1209 10:12:52.129636 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/065686b2-246c-4194-936f-b6e58ee9763b-utilities" (OuterVolumeSpecName: "utilities") pod "065686b2-246c-4194-936f-b6e58ee9763b" (UID: "065686b2-246c-4194-936f-b6e58ee9763b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:12:52 crc kubenswrapper[4824]: I1209 10:12:52.131313 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/065686b2-246c-4194-936f-b6e58ee9763b-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:52 crc kubenswrapper[4824]: I1209 10:12:52.140905 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/065686b2-246c-4194-936f-b6e58ee9763b-kube-api-access-bwfmq" (OuterVolumeSpecName: "kube-api-access-bwfmq") pod "065686b2-246c-4194-936f-b6e58ee9763b" (UID: "065686b2-246c-4194-936f-b6e58ee9763b"). InnerVolumeSpecName "kube-api-access-bwfmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:12:52 crc kubenswrapper[4824]: I1209 10:12:52.205937 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/065686b2-246c-4194-936f-b6e58ee9763b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "065686b2-246c-4194-936f-b6e58ee9763b" (UID: "065686b2-246c-4194-936f-b6e58ee9763b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:12:52 crc kubenswrapper[4824]: I1209 10:12:52.233601 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwfmq\" (UniqueName: \"kubernetes.io/projected/065686b2-246c-4194-936f-b6e58ee9763b-kube-api-access-bwfmq\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:52 crc kubenswrapper[4824]: I1209 10:12:52.233647 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/065686b2-246c-4194-936f-b6e58ee9763b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 10:12:52 crc kubenswrapper[4824]: I1209 10:12:52.306731 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6fffc89c95-wt6ms" event={"ID":"04bcf8d0-9439-4739-a503-e504651a80dd","Type":"ContainerStarted","Data":"ad73bbffb835406c8227acbaa03f03e4ddfc98ccf531b5e7a4e00a36acc7794a"} Dec 09 10:12:52 crc kubenswrapper[4824]: I1209 10:12:52.339367 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7466b579b4-qpdj6"] Dec 09 10:12:52 crc kubenswrapper[4824]: I1209 10:12:52.340116 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wx2h9" event={"ID":"065686b2-246c-4194-936f-b6e58ee9763b","Type":"ContainerDied","Data":"99ff906e3a59ca7f072add5e8e008da8d591b81df699822444eebd0ce7743340"} Dec 09 10:12:52 crc kubenswrapper[4824]: I1209 10:12:52.340178 4824 scope.go:117] "RemoveContainer" containerID="2d424024ddebda113c5fbcb03177ec44cb898168337f7d67047fc5844fbd2165" Dec 09 10:12:52 crc kubenswrapper[4824]: I1209 10:12:52.340385 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wx2h9" Dec 09 10:12:52 crc kubenswrapper[4824]: I1209 10:12:52.479596 4824 scope.go:117] "RemoveContainer" containerID="8d9fe5540c0ecc67d4aa35f991a15750e270b158cfe77e9cdbdfaebd836fa3c7" Dec 09 10:12:52 crc kubenswrapper[4824]: I1209 10:12:52.513496 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wx2h9"] Dec 09 10:12:52 crc kubenswrapper[4824]: I1209 10:12:52.523978 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wx2h9"] Dec 09 10:12:52 crc kubenswrapper[4824]: I1209 10:12:52.550326 4824 scope.go:117] "RemoveContainer" containerID="84d3866e15cc80d2a49e7eb9f9cf31c3dfc4bdf745a44a9943a62dbd7a5e0b4b" Dec 09 10:12:52 crc kubenswrapper[4824]: I1209 10:12:52.823309 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-79457998d5-fpwv7"] Dec 09 10:12:52 crc kubenswrapper[4824]: I1209 10:12:52.863689 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-tcz7b"] Dec 09 10:12:53 crc kubenswrapper[4824]: I1209 10:12:53.397914 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" event={"ID":"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8","Type":"ContainerStarted","Data":"c5c363f24710905cd95df44fce88db788c882b7a75ecece1756df8edb8e99e01"} Dec 09 10:12:53 crc kubenswrapper[4824]: I1209 10:12:53.408053 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7466b579b4-qpdj6" event={"ID":"0e74eb53-a9de-4679-99d1-7bb097c01d93","Type":"ContainerStarted","Data":"5852dc653bd94404eac8cf1fe5466d7c13ef236d394cdfabff8fc3878ba5fce6"} Dec 09 10:12:53 crc kubenswrapper[4824]: I1209 10:12:53.418953 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-79457998d5-fpwv7" event={"ID":"aeb18289-ca5e-45d5-8976-53d77b8c8ee9","Type":"ContainerStarted","Data":"fd16e607756ce62917f50427e7731ced31e470aa377d84ebc1c0b839ac791f45"} Dec 09 10:12:53 crc kubenswrapper[4824]: I1209 10:12:53.430943 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6fffc89c95-wt6ms" event={"ID":"04bcf8d0-9439-4739-a503-e504651a80dd","Type":"ContainerStarted","Data":"5b1cb02d98415e13ebd728dc83bdef09545646e7c9bbe6f2b829ceebe2c6ab0f"} Dec 09 10:12:53 crc kubenswrapper[4824]: I1209 10:12:53.431496 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-6fffc89c95-wt6ms" Dec 09 10:12:53 crc kubenswrapper[4824]: I1209 10:12:53.481013 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-6fffc89c95-wt6ms" podStartSLOduration=3.480983728 podStartE2EDuration="3.480983728s" podCreationTimestamp="2025-12-09 10:12:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:12:53.462355916 +0000 UTC m=+1529.796860583" watchObservedRunningTime="2025-12-09 10:12:53.480983728 +0000 UTC m=+1529.815488405" Dec 09 10:12:53 crc kubenswrapper[4824]: I1209 10:12:53.622605 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 09 10:12:53 crc kubenswrapper[4824]: E1209 10:12:53.813705 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod468592b2_a40d_47d8_aaf5_0b77b3f4e0a8.slice/crio-conmon-ccfba621e765d76b3511a4bae1bf6a4529bcb8b04d6255c797510fca2c5e97b8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod468592b2_a40d_47d8_aaf5_0b77b3f4e0a8.slice/crio-ccfba621e765d76b3511a4bae1bf6a4529bcb8b04d6255c797510fca2c5e97b8.scope\": RecentStats: unable to find data in memory cache]" Dec 09 10:12:54 crc kubenswrapper[4824]: I1209 10:12:54.103918 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="065686b2-246c-4194-936f-b6e58ee9763b" path="/var/lib/kubelet/pods/065686b2-246c-4194-936f-b6e58ee9763b/volumes" Dec 09 10:12:54 crc kubenswrapper[4824]: I1209 10:12:54.471512 4824 generic.go:334] "Generic (PLEG): container finished" podID="468592b2-a40d-47d8-aaf5-0b77b3f4e0a8" containerID="ccfba621e765d76b3511a4bae1bf6a4529bcb8b04d6255c797510fca2c5e97b8" exitCode=0 Dec 09 10:12:54 crc kubenswrapper[4824]: I1209 10:12:54.473305 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" event={"ID":"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8","Type":"ContainerDied","Data":"ccfba621e765d76b3511a4bae1bf6a4529bcb8b04d6255c797510fca2c5e97b8"} Dec 09 10:12:59 crc kubenswrapper[4824]: I1209 10:12:59.166002 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.205:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.004109 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-7d54cc949d-rzfx4"] Dec 09 10:13:01 crc kubenswrapper[4824]: E1209 10:13:01.005317 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="065686b2-246c-4194-936f-b6e58ee9763b" containerName="registry-server" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.005334 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="065686b2-246c-4194-936f-b6e58ee9763b" containerName="registry-server" Dec 09 10:13:01 crc kubenswrapper[4824]: E1209 10:13:01.005362 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="065686b2-246c-4194-936f-b6e58ee9763b" containerName="extract-utilities" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.005370 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="065686b2-246c-4194-936f-b6e58ee9763b" containerName="extract-utilities" Dec 09 10:13:01 crc kubenswrapper[4824]: E1209 10:13:01.005415 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="065686b2-246c-4194-936f-b6e58ee9763b" containerName="extract-content" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.005423 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="065686b2-246c-4194-936f-b6e58ee9763b" containerName="extract-content" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.005657 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="065686b2-246c-4194-936f-b6e58ee9763b" containerName="registry-server" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.006587 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7d54cc949d-rzfx4" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.027148 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7d54cc949d-rzfx4"] Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.101399 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-78488f59bc-grx2c"] Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.103964 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-78488f59bc-grx2c" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.117025 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-78488f59bc-grx2c"] Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.139608 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-69f49fcbcf-7bnk5"] Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.142673 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69f49fcbcf-7bnk5" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.706274 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpfng\" (UniqueName: \"kubernetes.io/projected/ddd52d5e-b6bd-403e-affe-c7d25aa4d556-kube-api-access-hpfng\") pod \"heat-cfnapi-78488f59bc-grx2c\" (UID: \"ddd52d5e-b6bd-403e-affe-c7d25aa4d556\") " pod="openstack/heat-cfnapi-78488f59bc-grx2c" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.706468 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rw8r\" (UniqueName: \"kubernetes.io/projected/d8154a7d-1dbc-4fa3-a862-140f2fd807b2-kube-api-access-7rw8r\") pod \"heat-engine-7d54cc949d-rzfx4\" (UID: \"d8154a7d-1dbc-4fa3-a862-140f2fd807b2\") " pod="openstack/heat-engine-7d54cc949d-rzfx4" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.707610 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8154a7d-1dbc-4fa3-a862-140f2fd807b2-combined-ca-bundle\") pod \"heat-engine-7d54cc949d-rzfx4\" (UID: \"d8154a7d-1dbc-4fa3-a862-140f2fd807b2\") " pod="openstack/heat-engine-7d54cc949d-rzfx4" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.708271 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8154a7d-1dbc-4fa3-a862-140f2fd807b2-config-data-custom\") pod \"heat-engine-7d54cc949d-rzfx4\" (UID: \"d8154a7d-1dbc-4fa3-a862-140f2fd807b2\") " pod="openstack/heat-engine-7d54cc949d-rzfx4" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.708478 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ddd52d5e-b6bd-403e-affe-c7d25aa4d556-config-data-custom\") pod \"heat-cfnapi-78488f59bc-grx2c\" (UID: \"ddd52d5e-b6bd-403e-affe-c7d25aa4d556\") " pod="openstack/heat-cfnapi-78488f59bc-grx2c" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.708563 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8154a7d-1dbc-4fa3-a862-140f2fd807b2-config-data\") pod \"heat-engine-7d54cc949d-rzfx4\" (UID: \"d8154a7d-1dbc-4fa3-a862-140f2fd807b2\") " pod="openstack/heat-engine-7d54cc949d-rzfx4" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.708616 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddd52d5e-b6bd-403e-affe-c7d25aa4d556-combined-ca-bundle\") pod \"heat-cfnapi-78488f59bc-grx2c\" (UID: \"ddd52d5e-b6bd-403e-affe-c7d25aa4d556\") " pod="openstack/heat-cfnapi-78488f59bc-grx2c" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.708711 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddd52d5e-b6bd-403e-affe-c7d25aa4d556-config-data\") pod \"heat-cfnapi-78488f59bc-grx2c\" (UID: \"ddd52d5e-b6bd-403e-affe-c7d25aa4d556\") " pod="openstack/heat-cfnapi-78488f59bc-grx2c" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.728951 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-69f49fcbcf-7bnk5"] Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.810704 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rw8r\" (UniqueName: \"kubernetes.io/projected/d8154a7d-1dbc-4fa3-a862-140f2fd807b2-kube-api-access-7rw8r\") pod \"heat-engine-7d54cc949d-rzfx4\" (UID: \"d8154a7d-1dbc-4fa3-a862-140f2fd807b2\") " pod="openstack/heat-engine-7d54cc949d-rzfx4" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.811222 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8154a7d-1dbc-4fa3-a862-140f2fd807b2-combined-ca-bundle\") pod \"heat-engine-7d54cc949d-rzfx4\" (UID: \"d8154a7d-1dbc-4fa3-a862-140f2fd807b2\") " pod="openstack/heat-engine-7d54cc949d-rzfx4" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.811362 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzpnl\" (UniqueName: \"kubernetes.io/projected/ec8287ef-1c22-4524-8a7b-c20d76430002-kube-api-access-gzpnl\") pod \"heat-api-69f49fcbcf-7bnk5\" (UID: \"ec8287ef-1c22-4524-8a7b-c20d76430002\") " pod="openstack/heat-api-69f49fcbcf-7bnk5" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.812769 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8287ef-1c22-4524-8a7b-c20d76430002-combined-ca-bundle\") pod \"heat-api-69f49fcbcf-7bnk5\" (UID: \"ec8287ef-1c22-4524-8a7b-c20d76430002\") " pod="openstack/heat-api-69f49fcbcf-7bnk5" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.814631 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ec8287ef-1c22-4524-8a7b-c20d76430002-config-data-custom\") pod \"heat-api-69f49fcbcf-7bnk5\" (UID: \"ec8287ef-1c22-4524-8a7b-c20d76430002\") " pod="openstack/heat-api-69f49fcbcf-7bnk5" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.814752 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8154a7d-1dbc-4fa3-a862-140f2fd807b2-config-data-custom\") pod \"heat-engine-7d54cc949d-rzfx4\" (UID: \"d8154a7d-1dbc-4fa3-a862-140f2fd807b2\") " pod="openstack/heat-engine-7d54cc949d-rzfx4" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.814951 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ddd52d5e-b6bd-403e-affe-c7d25aa4d556-config-data-custom\") pod \"heat-cfnapi-78488f59bc-grx2c\" (UID: \"ddd52d5e-b6bd-403e-affe-c7d25aa4d556\") " pod="openstack/heat-cfnapi-78488f59bc-grx2c" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.815027 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8154a7d-1dbc-4fa3-a862-140f2fd807b2-config-data\") pod \"heat-engine-7d54cc949d-rzfx4\" (UID: \"d8154a7d-1dbc-4fa3-a862-140f2fd807b2\") " pod="openstack/heat-engine-7d54cc949d-rzfx4" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.815068 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddd52d5e-b6bd-403e-affe-c7d25aa4d556-combined-ca-bundle\") pod \"heat-cfnapi-78488f59bc-grx2c\" (UID: \"ddd52d5e-b6bd-403e-affe-c7d25aa4d556\") " pod="openstack/heat-cfnapi-78488f59bc-grx2c" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.815169 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddd52d5e-b6bd-403e-affe-c7d25aa4d556-config-data\") pod \"heat-cfnapi-78488f59bc-grx2c\" (UID: \"ddd52d5e-b6bd-403e-affe-c7d25aa4d556\") " pod="openstack/heat-cfnapi-78488f59bc-grx2c" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.815232 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpfng\" (UniqueName: \"kubernetes.io/projected/ddd52d5e-b6bd-403e-affe-c7d25aa4d556-kube-api-access-hpfng\") pod \"heat-cfnapi-78488f59bc-grx2c\" (UID: \"ddd52d5e-b6bd-403e-affe-c7d25aa4d556\") " pod="openstack/heat-cfnapi-78488f59bc-grx2c" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.815277 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8287ef-1c22-4524-8a7b-c20d76430002-config-data\") pod \"heat-api-69f49fcbcf-7bnk5\" (UID: \"ec8287ef-1c22-4524-8a7b-c20d76430002\") " pod="openstack/heat-api-69f49fcbcf-7bnk5" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.823908 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8154a7d-1dbc-4fa3-a862-140f2fd807b2-config-data\") pod \"heat-engine-7d54cc949d-rzfx4\" (UID: \"d8154a7d-1dbc-4fa3-a862-140f2fd807b2\") " pod="openstack/heat-engine-7d54cc949d-rzfx4" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.831767 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8154a7d-1dbc-4fa3-a862-140f2fd807b2-combined-ca-bundle\") pod \"heat-engine-7d54cc949d-rzfx4\" (UID: \"d8154a7d-1dbc-4fa3-a862-140f2fd807b2\") " pod="openstack/heat-engine-7d54cc949d-rzfx4" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.832434 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8154a7d-1dbc-4fa3-a862-140f2fd807b2-config-data-custom\") pod \"heat-engine-7d54cc949d-rzfx4\" (UID: \"d8154a7d-1dbc-4fa3-a862-140f2fd807b2\") " pod="openstack/heat-engine-7d54cc949d-rzfx4" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.832731 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddd52d5e-b6bd-403e-affe-c7d25aa4d556-combined-ca-bundle\") pod \"heat-cfnapi-78488f59bc-grx2c\" (UID: \"ddd52d5e-b6bd-403e-affe-c7d25aa4d556\") " pod="openstack/heat-cfnapi-78488f59bc-grx2c" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.833628 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddd52d5e-b6bd-403e-affe-c7d25aa4d556-config-data\") pod \"heat-cfnapi-78488f59bc-grx2c\" (UID: \"ddd52d5e-b6bd-403e-affe-c7d25aa4d556\") " pod="openstack/heat-cfnapi-78488f59bc-grx2c" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.834739 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ddd52d5e-b6bd-403e-affe-c7d25aa4d556-config-data-custom\") pod \"heat-cfnapi-78488f59bc-grx2c\" (UID: \"ddd52d5e-b6bd-403e-affe-c7d25aa4d556\") " pod="openstack/heat-cfnapi-78488f59bc-grx2c" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.834885 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rw8r\" (UniqueName: \"kubernetes.io/projected/d8154a7d-1dbc-4fa3-a862-140f2fd807b2-kube-api-access-7rw8r\") pod \"heat-engine-7d54cc949d-rzfx4\" (UID: \"d8154a7d-1dbc-4fa3-a862-140f2fd807b2\") " pod="openstack/heat-engine-7d54cc949d-rzfx4" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.841732 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpfng\" (UniqueName: \"kubernetes.io/projected/ddd52d5e-b6bd-403e-affe-c7d25aa4d556-kube-api-access-hpfng\") pod \"heat-cfnapi-78488f59bc-grx2c\" (UID: \"ddd52d5e-b6bd-403e-affe-c7d25aa4d556\") " pod="openstack/heat-cfnapi-78488f59bc-grx2c" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.917940 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzpnl\" (UniqueName: \"kubernetes.io/projected/ec8287ef-1c22-4524-8a7b-c20d76430002-kube-api-access-gzpnl\") pod \"heat-api-69f49fcbcf-7bnk5\" (UID: \"ec8287ef-1c22-4524-8a7b-c20d76430002\") " pod="openstack/heat-api-69f49fcbcf-7bnk5" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.918308 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8287ef-1c22-4524-8a7b-c20d76430002-combined-ca-bundle\") pod \"heat-api-69f49fcbcf-7bnk5\" (UID: \"ec8287ef-1c22-4524-8a7b-c20d76430002\") " pod="openstack/heat-api-69f49fcbcf-7bnk5" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.918374 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ec8287ef-1c22-4524-8a7b-c20d76430002-config-data-custom\") pod \"heat-api-69f49fcbcf-7bnk5\" (UID: \"ec8287ef-1c22-4524-8a7b-c20d76430002\") " pod="openstack/heat-api-69f49fcbcf-7bnk5" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.918589 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8287ef-1c22-4524-8a7b-c20d76430002-config-data\") pod \"heat-api-69f49fcbcf-7bnk5\" (UID: \"ec8287ef-1c22-4524-8a7b-c20d76430002\") " pod="openstack/heat-api-69f49fcbcf-7bnk5" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.932774 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8287ef-1c22-4524-8a7b-c20d76430002-combined-ca-bundle\") pod \"heat-api-69f49fcbcf-7bnk5\" (UID: \"ec8287ef-1c22-4524-8a7b-c20d76430002\") " pod="openstack/heat-api-69f49fcbcf-7bnk5" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.933745 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ec8287ef-1c22-4524-8a7b-c20d76430002-config-data-custom\") pod \"heat-api-69f49fcbcf-7bnk5\" (UID: \"ec8287ef-1c22-4524-8a7b-c20d76430002\") " pod="openstack/heat-api-69f49fcbcf-7bnk5" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.935954 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8287ef-1c22-4524-8a7b-c20d76430002-config-data\") pod \"heat-api-69f49fcbcf-7bnk5\" (UID: \"ec8287ef-1c22-4524-8a7b-c20d76430002\") " pod="openstack/heat-api-69f49fcbcf-7bnk5" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.943720 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzpnl\" (UniqueName: \"kubernetes.io/projected/ec8287ef-1c22-4524-8a7b-c20d76430002-kube-api-access-gzpnl\") pod \"heat-api-69f49fcbcf-7bnk5\" (UID: \"ec8287ef-1c22-4524-8a7b-c20d76430002\") " pod="openstack/heat-api-69f49fcbcf-7bnk5" Dec 09 10:13:01 crc kubenswrapper[4824]: I1209 10:13:01.946576 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7d54cc949d-rzfx4" Dec 09 10:13:02 crc kubenswrapper[4824]: I1209 10:13:02.038807 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-78488f59bc-grx2c" Dec 09 10:13:02 crc kubenswrapper[4824]: I1209 10:13:02.058938 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69f49fcbcf-7bnk5" Dec 09 10:13:02 crc kubenswrapper[4824]: I1209 10:13:02.145533 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:13:02 crc kubenswrapper[4824]: I1209 10:13:02.145991 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="188ccefd-703d-4857-a005-70c0ce68d09c" containerName="ceilometer-central-agent" containerID="cri-o://86c269b7bff445ae59c766a934fd866cc747ab648a33b2f9c4e8ec19aef44a93" gracePeriod=30 Dec 09 10:13:02 crc kubenswrapper[4824]: I1209 10:13:02.146025 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="188ccefd-703d-4857-a005-70c0ce68d09c" containerName="proxy-httpd" containerID="cri-o://dbc7f9130eac008823602e68826e11f4cd01648aa902f777ec75e36f8d466412" gracePeriod=30 Dec 09 10:13:02 crc kubenswrapper[4824]: I1209 10:13:02.146144 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="188ccefd-703d-4857-a005-70c0ce68d09c" containerName="sg-core" containerID="cri-o://060ba5d4bcaabe29261b98d0afd9b5e28d1447b92d0016faa1fabc5823425c22" gracePeriod=30 Dec 09 10:13:02 crc kubenswrapper[4824]: I1209 10:13:02.146184 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="188ccefd-703d-4857-a005-70c0ce68d09c" containerName="ceilometer-notification-agent" containerID="cri-o://fe26ca2e87cd228b8c6a6502631ef82a32112cc0374160767f4acd98b820dd7b" gracePeriod=30 Dec 09 10:13:02 crc kubenswrapper[4824]: I1209 10:13:02.166554 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="188ccefd-703d-4857-a005-70c0ce68d09c" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.206:3000/\": EOF" Dec 09 10:13:02 crc kubenswrapper[4824]: I1209 10:13:02.844013 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-85dcdb56bc-4xfwk"] Dec 09 10:13:02 crc kubenswrapper[4824]: I1209 10:13:02.847516 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:02 crc kubenswrapper[4824]: I1209 10:13:02.851982 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 09 10:13:02 crc kubenswrapper[4824]: I1209 10:13:02.852048 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 09 10:13:02 crc kubenswrapper[4824]: I1209 10:13:02.855316 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 09 10:13:02 crc kubenswrapper[4824]: I1209 10:13:02.872311 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-85dcdb56bc-4xfwk"] Dec 09 10:13:02 crc kubenswrapper[4824]: I1209 10:13:02.960189 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bl8x\" (UniqueName: \"kubernetes.io/projected/574c42ce-f75f-4dab-bd77-916b9a4d2c6d-kube-api-access-4bl8x\") pod \"swift-proxy-85dcdb56bc-4xfwk\" (UID: \"574c42ce-f75f-4dab-bd77-916b9a4d2c6d\") " pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:02 crc kubenswrapper[4824]: I1209 10:13:02.960437 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/574c42ce-f75f-4dab-bd77-916b9a4d2c6d-internal-tls-certs\") pod \"swift-proxy-85dcdb56bc-4xfwk\" (UID: \"574c42ce-f75f-4dab-bd77-916b9a4d2c6d\") " pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:02 crc kubenswrapper[4824]: I1209 10:13:02.960551 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/574c42ce-f75f-4dab-bd77-916b9a4d2c6d-public-tls-certs\") pod \"swift-proxy-85dcdb56bc-4xfwk\" (UID: \"574c42ce-f75f-4dab-bd77-916b9a4d2c6d\") " pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:02 crc kubenswrapper[4824]: I1209 10:13:02.960607 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/574c42ce-f75f-4dab-bd77-916b9a4d2c6d-combined-ca-bundle\") pod \"swift-proxy-85dcdb56bc-4xfwk\" (UID: \"574c42ce-f75f-4dab-bd77-916b9a4d2c6d\") " pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:02 crc kubenswrapper[4824]: I1209 10:13:02.960912 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/574c42ce-f75f-4dab-bd77-916b9a4d2c6d-etc-swift\") pod \"swift-proxy-85dcdb56bc-4xfwk\" (UID: \"574c42ce-f75f-4dab-bd77-916b9a4d2c6d\") " pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:02 crc kubenswrapper[4824]: I1209 10:13:02.961140 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/574c42ce-f75f-4dab-bd77-916b9a4d2c6d-config-data\") pod \"swift-proxy-85dcdb56bc-4xfwk\" (UID: \"574c42ce-f75f-4dab-bd77-916b9a4d2c6d\") " pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:02 crc kubenswrapper[4824]: I1209 10:13:02.961401 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/574c42ce-f75f-4dab-bd77-916b9a4d2c6d-run-httpd\") pod \"swift-proxy-85dcdb56bc-4xfwk\" (UID: \"574c42ce-f75f-4dab-bd77-916b9a4d2c6d\") " pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:02 crc kubenswrapper[4824]: I1209 10:13:02.961768 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/574c42ce-f75f-4dab-bd77-916b9a4d2c6d-log-httpd\") pod \"swift-proxy-85dcdb56bc-4xfwk\" (UID: \"574c42ce-f75f-4dab-bd77-916b9a4d2c6d\") " pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.039135 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="188ccefd-703d-4857-a005-70c0ce68d09c" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.206:3000/\": dial tcp 10.217.0.206:3000: connect: connection refused" Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.089519 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bl8x\" (UniqueName: \"kubernetes.io/projected/574c42ce-f75f-4dab-bd77-916b9a4d2c6d-kube-api-access-4bl8x\") pod \"swift-proxy-85dcdb56bc-4xfwk\" (UID: \"574c42ce-f75f-4dab-bd77-916b9a4d2c6d\") " pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.089578 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/574c42ce-f75f-4dab-bd77-916b9a4d2c6d-internal-tls-certs\") pod \"swift-proxy-85dcdb56bc-4xfwk\" (UID: \"574c42ce-f75f-4dab-bd77-916b9a4d2c6d\") " pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.089617 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/574c42ce-f75f-4dab-bd77-916b9a4d2c6d-public-tls-certs\") pod \"swift-proxy-85dcdb56bc-4xfwk\" (UID: \"574c42ce-f75f-4dab-bd77-916b9a4d2c6d\") " pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.089645 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/574c42ce-f75f-4dab-bd77-916b9a4d2c6d-combined-ca-bundle\") pod \"swift-proxy-85dcdb56bc-4xfwk\" (UID: \"574c42ce-f75f-4dab-bd77-916b9a4d2c6d\") " pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.089801 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/574c42ce-f75f-4dab-bd77-916b9a4d2c6d-etc-swift\") pod \"swift-proxy-85dcdb56bc-4xfwk\" (UID: \"574c42ce-f75f-4dab-bd77-916b9a4d2c6d\") " pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.089880 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/574c42ce-f75f-4dab-bd77-916b9a4d2c6d-config-data\") pod \"swift-proxy-85dcdb56bc-4xfwk\" (UID: \"574c42ce-f75f-4dab-bd77-916b9a4d2c6d\") " pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.090031 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/574c42ce-f75f-4dab-bd77-916b9a4d2c6d-run-httpd\") pod \"swift-proxy-85dcdb56bc-4xfwk\" (UID: \"574c42ce-f75f-4dab-bd77-916b9a4d2c6d\") " pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.090231 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/574c42ce-f75f-4dab-bd77-916b9a4d2c6d-log-httpd\") pod \"swift-proxy-85dcdb56bc-4xfwk\" (UID: \"574c42ce-f75f-4dab-bd77-916b9a4d2c6d\") " pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.092536 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/574c42ce-f75f-4dab-bd77-916b9a4d2c6d-run-httpd\") pod \"swift-proxy-85dcdb56bc-4xfwk\" (UID: \"574c42ce-f75f-4dab-bd77-916b9a4d2c6d\") " pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.093539 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/574c42ce-f75f-4dab-bd77-916b9a4d2c6d-log-httpd\") pod \"swift-proxy-85dcdb56bc-4xfwk\" (UID: \"574c42ce-f75f-4dab-bd77-916b9a4d2c6d\") " pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.099945 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/574c42ce-f75f-4dab-bd77-916b9a4d2c6d-combined-ca-bundle\") pod \"swift-proxy-85dcdb56bc-4xfwk\" (UID: \"574c42ce-f75f-4dab-bd77-916b9a4d2c6d\") " pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.101054 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/574c42ce-f75f-4dab-bd77-916b9a4d2c6d-config-data\") pod \"swift-proxy-85dcdb56bc-4xfwk\" (UID: \"574c42ce-f75f-4dab-bd77-916b9a4d2c6d\") " pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.101934 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/574c42ce-f75f-4dab-bd77-916b9a4d2c6d-public-tls-certs\") pod \"swift-proxy-85dcdb56bc-4xfwk\" (UID: \"574c42ce-f75f-4dab-bd77-916b9a4d2c6d\") " pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.102289 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/574c42ce-f75f-4dab-bd77-916b9a4d2c6d-etc-swift\") pod \"swift-proxy-85dcdb56bc-4xfwk\" (UID: \"574c42ce-f75f-4dab-bd77-916b9a4d2c6d\") " pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.121769 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/574c42ce-f75f-4dab-bd77-916b9a4d2c6d-internal-tls-certs\") pod \"swift-proxy-85dcdb56bc-4xfwk\" (UID: \"574c42ce-f75f-4dab-bd77-916b9a4d2c6d\") " pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.131444 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bl8x\" (UniqueName: \"kubernetes.io/projected/574c42ce-f75f-4dab-bd77-916b9a4d2c6d-kube-api-access-4bl8x\") pod \"swift-proxy-85dcdb56bc-4xfwk\" (UID: \"574c42ce-f75f-4dab-bd77-916b9a4d2c6d\") " pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.168673 4824 generic.go:334] "Generic (PLEG): container finished" podID="6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b" containerID="09192e79e2b24b4ba4ff045af664c1a16f27e0e694d1b16486716f09b6c84652" exitCode=137 Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.168753 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b","Type":"ContainerDied","Data":"09192e79e2b24b4ba4ff045af664c1a16f27e0e694d1b16486716f09b6c84652"} Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.183679 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.188131 4824 generic.go:334] "Generic (PLEG): container finished" podID="188ccefd-703d-4857-a005-70c0ce68d09c" containerID="dbc7f9130eac008823602e68826e11f4cd01648aa902f777ec75e36f8d466412" exitCode=0 Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.188175 4824 generic.go:334] "Generic (PLEG): container finished" podID="188ccefd-703d-4857-a005-70c0ce68d09c" containerID="060ba5d4bcaabe29261b98d0afd9b5e28d1447b92d0016faa1fabc5823425c22" exitCode=2 Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.188947 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"188ccefd-703d-4857-a005-70c0ce68d09c","Type":"ContainerDied","Data":"dbc7f9130eac008823602e68826e11f4cd01648aa902f777ec75e36f8d466412"} Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.189006 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"188ccefd-703d-4857-a005-70c0ce68d09c","Type":"ContainerDied","Data":"060ba5d4bcaabe29261b98d0afd9b5e28d1447b92d0016faa1fabc5823425c22"} Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.189020 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"188ccefd-703d-4857-a005-70c0ce68d09c","Type":"ContainerDied","Data":"86c269b7bff445ae59c766a934fd866cc747ab648a33b2f9c4e8ec19aef44a93"} Dec 09 10:13:03 crc kubenswrapper[4824]: I1209 10:13:03.194987 4824 generic.go:334] "Generic (PLEG): container finished" podID="188ccefd-703d-4857-a005-70c0ce68d09c" containerID="86c269b7bff445ae59c766a934fd866cc747ab648a33b2f9c4e8ec19aef44a93" exitCode=0 Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.124423 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.205:8776/healthcheck\": dial tcp 10.217.0.205:8776: connect: connection refused" Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.719416 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-79457998d5-fpwv7"] Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.732652 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7466b579b4-qpdj6"] Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.784048 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-98f88b69-9xf7z"] Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.786373 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-98f88b69-9xf7z" Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.789150 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.789613 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.837206 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-98f88b69-9xf7z"] Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.850685 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-ff546578f-d6scd"] Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.854921 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-ff546578f-d6scd" Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.857577 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.857705 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.864125 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-ff546578f-d6scd"] Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.961416 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-combined-ca-bundle\") pod \"heat-api-98f88b69-9xf7z\" (UID: \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\") " pod="openstack/heat-api-98f88b69-9xf7z" Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.961527 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-config-data\") pod \"heat-cfnapi-ff546578f-d6scd\" (UID: \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\") " pod="openstack/heat-cfnapi-ff546578f-d6scd" Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.961559 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-public-tls-certs\") pod \"heat-cfnapi-ff546578f-d6scd\" (UID: \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\") " pod="openstack/heat-cfnapi-ff546578f-d6scd" Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.961595 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6skhn\" (UniqueName: \"kubernetes.io/projected/14358ed8-0c10-48f7-a1b1-37e75e7a6400-kube-api-access-6skhn\") pod \"heat-cfnapi-ff546578f-d6scd\" (UID: \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\") " pod="openstack/heat-cfnapi-ff546578f-d6scd" Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.961821 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-config-data-custom\") pod \"heat-api-98f88b69-9xf7z\" (UID: \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\") " pod="openstack/heat-api-98f88b69-9xf7z" Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.961983 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8lpt\" (UniqueName: \"kubernetes.io/projected/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-kube-api-access-f8lpt\") pod \"heat-api-98f88b69-9xf7z\" (UID: \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\") " pod="openstack/heat-api-98f88b69-9xf7z" Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.966977 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-internal-tls-certs\") pod \"heat-api-98f88b69-9xf7z\" (UID: \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\") " pod="openstack/heat-api-98f88b69-9xf7z" Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.967023 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-combined-ca-bundle\") pod \"heat-cfnapi-ff546578f-d6scd\" (UID: \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\") " pod="openstack/heat-cfnapi-ff546578f-d6scd" Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.967144 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-config-data\") pod \"heat-api-98f88b69-9xf7z\" (UID: \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\") " pod="openstack/heat-api-98f88b69-9xf7z" Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.967225 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-public-tls-certs\") pod \"heat-api-98f88b69-9xf7z\" (UID: \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\") " pod="openstack/heat-api-98f88b69-9xf7z" Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.967346 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-config-data-custom\") pod \"heat-cfnapi-ff546578f-d6scd\" (UID: \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\") " pod="openstack/heat-cfnapi-ff546578f-d6scd" Dec 09 10:13:04 crc kubenswrapper[4824]: I1209 10:13:04.968464 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-internal-tls-certs\") pod \"heat-cfnapi-ff546578f-d6scd\" (UID: \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\") " pod="openstack/heat-cfnapi-ff546578f-d6scd" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.071244 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-public-tls-certs\") pod \"heat-api-98f88b69-9xf7z\" (UID: \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\") " pod="openstack/heat-api-98f88b69-9xf7z" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.072002 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-config-data-custom\") pod \"heat-cfnapi-ff546578f-d6scd\" (UID: \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\") " pod="openstack/heat-cfnapi-ff546578f-d6scd" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.072505 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-internal-tls-certs\") pod \"heat-cfnapi-ff546578f-d6scd\" (UID: \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\") " pod="openstack/heat-cfnapi-ff546578f-d6scd" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.072672 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-combined-ca-bundle\") pod \"heat-api-98f88b69-9xf7z\" (UID: \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\") " pod="openstack/heat-api-98f88b69-9xf7z" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.072843 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-config-data\") pod \"heat-cfnapi-ff546578f-d6scd\" (UID: \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\") " pod="openstack/heat-cfnapi-ff546578f-d6scd" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.073159 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-public-tls-certs\") pod \"heat-cfnapi-ff546578f-d6scd\" (UID: \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\") " pod="openstack/heat-cfnapi-ff546578f-d6scd" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.073448 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6skhn\" (UniqueName: \"kubernetes.io/projected/14358ed8-0c10-48f7-a1b1-37e75e7a6400-kube-api-access-6skhn\") pod \"heat-cfnapi-ff546578f-d6scd\" (UID: \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\") " pod="openstack/heat-cfnapi-ff546578f-d6scd" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.073610 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-config-data-custom\") pod \"heat-api-98f88b69-9xf7z\" (UID: \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\") " pod="openstack/heat-api-98f88b69-9xf7z" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.073902 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8lpt\" (UniqueName: \"kubernetes.io/projected/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-kube-api-access-f8lpt\") pod \"heat-api-98f88b69-9xf7z\" (UID: \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\") " pod="openstack/heat-api-98f88b69-9xf7z" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.074304 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-internal-tls-certs\") pod \"heat-api-98f88b69-9xf7z\" (UID: \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\") " pod="openstack/heat-api-98f88b69-9xf7z" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.074517 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-combined-ca-bundle\") pod \"heat-cfnapi-ff546578f-d6scd\" (UID: \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\") " pod="openstack/heat-cfnapi-ff546578f-d6scd" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.074752 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-config-data\") pod \"heat-api-98f88b69-9xf7z\" (UID: \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\") " pod="openstack/heat-api-98f88b69-9xf7z" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.087411 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-combined-ca-bundle\") pod \"heat-cfnapi-ff546578f-d6scd\" (UID: \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\") " pod="openstack/heat-cfnapi-ff546578f-d6scd" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.087466 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-combined-ca-bundle\") pod \"heat-api-98f88b69-9xf7z\" (UID: \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\") " pod="openstack/heat-api-98f88b69-9xf7z" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.088210 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-public-tls-certs\") pod \"heat-api-98f88b69-9xf7z\" (UID: \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\") " pod="openstack/heat-api-98f88b69-9xf7z" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.089031 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-internal-tls-certs\") pod \"heat-cfnapi-ff546578f-d6scd\" (UID: \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\") " pod="openstack/heat-cfnapi-ff546578f-d6scd" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.091902 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-config-data\") pod \"heat-api-98f88b69-9xf7z\" (UID: \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\") " pod="openstack/heat-api-98f88b69-9xf7z" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.091899 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-config-data\") pod \"heat-cfnapi-ff546578f-d6scd\" (UID: \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\") " pod="openstack/heat-cfnapi-ff546578f-d6scd" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.093600 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-config-data-custom\") pod \"heat-api-98f88b69-9xf7z\" (UID: \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\") " pod="openstack/heat-api-98f88b69-9xf7z" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.094842 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-internal-tls-certs\") pod \"heat-api-98f88b69-9xf7z\" (UID: \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\") " pod="openstack/heat-api-98f88b69-9xf7z" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.094860 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-config-data-custom\") pod \"heat-cfnapi-ff546578f-d6scd\" (UID: \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\") " pod="openstack/heat-cfnapi-ff546578f-d6scd" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.104748 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-public-tls-certs\") pod \"heat-cfnapi-ff546578f-d6scd\" (UID: \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\") " pod="openstack/heat-cfnapi-ff546578f-d6scd" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.105485 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6skhn\" (UniqueName: \"kubernetes.io/projected/14358ed8-0c10-48f7-a1b1-37e75e7a6400-kube-api-access-6skhn\") pod \"heat-cfnapi-ff546578f-d6scd\" (UID: \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\") " pod="openstack/heat-cfnapi-ff546578f-d6scd" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.111726 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8lpt\" (UniqueName: \"kubernetes.io/projected/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-kube-api-access-f8lpt\") pod \"heat-api-98f88b69-9xf7z\" (UID: \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\") " pod="openstack/heat-api-98f88b69-9xf7z" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.133656 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-98f88b69-9xf7z" Dec 09 10:13:05 crc kubenswrapper[4824]: I1209 10:13:05.179021 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-ff546578f-d6scd" Dec 09 10:13:06 crc kubenswrapper[4824]: I1209 10:13:06.523216 4824 generic.go:334] "Generic (PLEG): container finished" podID="188ccefd-703d-4857-a005-70c0ce68d09c" containerID="fe26ca2e87cd228b8c6a6502631ef82a32112cc0374160767f4acd98b820dd7b" exitCode=0 Dec 09 10:13:06 crc kubenswrapper[4824]: I1209 10:13:06.523446 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"188ccefd-703d-4857-a005-70c0ce68d09c","Type":"ContainerDied","Data":"fe26ca2e87cd228b8c6a6502631ef82a32112cc0374160767f4acd98b820dd7b"} Dec 09 10:13:08 crc kubenswrapper[4824]: E1209 10:13:08.087556 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified" Dec 09 10:13:08 crc kubenswrapper[4824]: E1209 10:13:08.088217 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstackclient,Image:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,Command:[/bin/sleep],Args:[infinity],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n54dh5d5hf8h64dh679h5ddh556h98h8bh557h647h66ch644h664h679h555h5h59fh594h545h76hd7h5c5h668h9dhb6h698h55dh7dh98h75h5dbq,ValueFrom:nil,},EnvVar{Name:OS_CLOUD,Value:default,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_CA_CERT,Value:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_HOST,Value:metric-storage-prometheus.openstack.svc,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_PORT,Value:9090,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openstack-config,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/cloudrc,SubPath:cloudrc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-scxsc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42401,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42401,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstackclient_openstack(a00596c6-582b-4ec7-9920-6dddc8197001): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:13:08 crc kubenswrapper[4824]: E1209 10:13:08.092997 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstackclient" podUID="a00596c6-582b-4ec7-9920-6dddc8197001" Dec 09 10:13:08 crc kubenswrapper[4824]: E1209 10:13:08.527566 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified" Dec 09 10:13:08 crc kubenswrapper[4824]: E1209 10:13:08.527883 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-cfnapi,Image:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_httpd_setup && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n698hfdh98h575h58bh68bh658h584hdbh5b6h685h5cch558h694hb6hbh597h69h55fhbhbbh5f4h5cfh57h8h95h699hc4h686h57ch88h54bq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:heat-cfnapi-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-custom,ReadOnly:true,MountPath:/etc/heat/heat.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tlg58,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthcheck,Port:{0 8000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:10,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthcheck,Port:{0 8000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:10,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-cfnapi-7466b579b4-qpdj6_openstack(0e74eb53-a9de-4679-99d1-7bb097c01d93): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:13:08 crc kubenswrapper[4824]: E1209 10:13:08.529836 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-cfnapi-7466b579b4-qpdj6" podUID="0e74eb53-a9de-4679-99d1-7bb097c01d93" Dec 09 10:13:08 crc kubenswrapper[4824]: E1209 10:13:08.638875 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified\\\"\"" pod="openstack/openstackclient" podUID="a00596c6-582b-4ec7-9920-6dddc8197001" Dec 09 10:13:10 crc kubenswrapper[4824]: E1209 10:13:10.177878 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-api:current-podified" Dec 09 10:13:10 crc kubenswrapper[4824]: E1209 10:13:10.178775 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-api,Image:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_httpd_setup && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n674h66bh675h5b8h5dbh58ch58ch59h9h7fhdh568h5cch675h547h5f4h5h679hbfhf5h77h577h675h56bh674h546h589h8fh58h57h5c7h5fdq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:heat-api-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-custom,ReadOnly:true,MountPath:/etc/heat/heat.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vdlzz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthcheck,Port:{0 8004 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:10,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthcheck,Port:{0 8004 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:10,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-api-79457998d5-fpwv7_openstack(aeb18289-ca5e-45d5-8976-53d77b8c8ee9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:13:10 crc kubenswrapper[4824]: E1209 10:13:10.180252 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-api-79457998d5-fpwv7" podUID="aeb18289-ca5e-45d5-8976-53d77b8c8ee9" Dec 09 10:13:10 crc kubenswrapper[4824]: I1209 10:13:10.703501 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 10:13:10 crc kubenswrapper[4824]: I1209 10:13:10.772355 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:13:10 crc kubenswrapper[4824]: I1209 10:13:10.776002 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"188ccefd-703d-4857-a005-70c0ce68d09c","Type":"ContainerDied","Data":"619d823c4dcabead9a5845864f1822ecce507d47502f5efe978f61491f5d3bfb"} Dec 09 10:13:10 crc kubenswrapper[4824]: I1209 10:13:10.776059 4824 scope.go:117] "RemoveContainer" containerID="dbc7f9130eac008823602e68826e11f4cd01648aa902f777ec75e36f8d466412" Dec 09 10:13:10 crc kubenswrapper[4824]: I1209 10:13:10.776566 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7466b579b4-qpdj6" Dec 09 10:13:10 crc kubenswrapper[4824]: I1209 10:13:10.806583 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7466b579b4-qpdj6" event={"ID":"0e74eb53-a9de-4679-99d1-7bb097c01d93","Type":"ContainerDied","Data":"5852dc653bd94404eac8cf1fe5466d7c13ef236d394cdfabff8fc3878ba5fce6"} Dec 09 10:13:10 crc kubenswrapper[4824]: I1209 10:13:10.843578 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-etc-machine-id\") pod \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " Dec 09 10:13:10 crc kubenswrapper[4824]: I1209 10:13:10.843722 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-scripts\") pod \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " Dec 09 10:13:10 crc kubenswrapper[4824]: I1209 10:13:10.843770 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-config-data\") pod \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " Dec 09 10:13:10 crc kubenswrapper[4824]: I1209 10:13:10.843845 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-config-data-custom\") pod \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " Dec 09 10:13:10 crc kubenswrapper[4824]: I1209 10:13:10.844041 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcf5b\" (UniqueName: \"kubernetes.io/projected/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-kube-api-access-gcf5b\") pod \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " Dec 09 10:13:10 crc kubenswrapper[4824]: I1209 10:13:10.844205 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-combined-ca-bundle\") pod \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " Dec 09 10:13:10 crc kubenswrapper[4824]: I1209 10:13:10.844269 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-logs\") pod \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\" (UID: \"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b\") " Dec 09 10:13:10 crc kubenswrapper[4824]: I1209 10:13:10.846915 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b" (UID: "6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 10:13:10 crc kubenswrapper[4824]: I1209 10:13:10.848226 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-logs" (OuterVolumeSpecName: "logs") pod "6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b" (UID: "6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:13:10 crc kubenswrapper[4824]: I1209 10:13:10.875464 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b","Type":"ContainerDied","Data":"ea28c43ec2abec5438ddaa9acd743f758507ffd3669148871909fd04484e3cf0"} Dec 09 10:13:10 crc kubenswrapper[4824]: I1209 10:13:10.902913 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 10:13:10 crc kubenswrapper[4824]: I1209 10:13:10.987049 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-kube-api-access-gcf5b" (OuterVolumeSpecName: "kube-api-access-gcf5b") pod "6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b" (UID: "6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b"). InnerVolumeSpecName "kube-api-access-gcf5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.231719 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-scripts" (OuterVolumeSpecName: "scripts") pod "6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b" (UID: "6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.239121 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b" (UID: "6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.277532 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-logs\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.277861 4824 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.290108 4824 scope.go:117] "RemoveContainer" containerID="060ba5d4bcaabe29261b98d0afd9b5e28d1447b92d0016faa1fabc5823425c22" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.379131 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e74eb53-a9de-4679-99d1-7bb097c01d93-config-data\") pod \"0e74eb53-a9de-4679-99d1-7bb097c01d93\" (UID: \"0e74eb53-a9de-4679-99d1-7bb097c01d93\") " Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.382154 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/188ccefd-703d-4857-a005-70c0ce68d09c-sg-core-conf-yaml\") pod \"188ccefd-703d-4857-a005-70c0ce68d09c\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.382700 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/188ccefd-703d-4857-a005-70c0ce68d09c-run-httpd\") pod \"188ccefd-703d-4857-a005-70c0ce68d09c\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.382754 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e74eb53-a9de-4679-99d1-7bb097c01d93-combined-ca-bundle\") pod \"0e74eb53-a9de-4679-99d1-7bb097c01d93\" (UID: \"0e74eb53-a9de-4679-99d1-7bb097c01d93\") " Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.382865 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188ccefd-703d-4857-a005-70c0ce68d09c-combined-ca-bundle\") pod \"188ccefd-703d-4857-a005-70c0ce68d09c\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.382952 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e74eb53-a9de-4679-99d1-7bb097c01d93-config-data-custom\") pod \"0e74eb53-a9de-4679-99d1-7bb097c01d93\" (UID: \"0e74eb53-a9de-4679-99d1-7bb097c01d93\") " Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.383119 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188ccefd-703d-4857-a005-70c0ce68d09c-config-data\") pod \"188ccefd-703d-4857-a005-70c0ce68d09c\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.383196 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/188ccefd-703d-4857-a005-70c0ce68d09c-log-httpd\") pod \"188ccefd-703d-4857-a005-70c0ce68d09c\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.383234 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlg58\" (UniqueName: \"kubernetes.io/projected/0e74eb53-a9de-4679-99d1-7bb097c01d93-kube-api-access-tlg58\") pod \"0e74eb53-a9de-4679-99d1-7bb097c01d93\" (UID: \"0e74eb53-a9de-4679-99d1-7bb097c01d93\") " Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.383303 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7k6l\" (UniqueName: \"kubernetes.io/projected/188ccefd-703d-4857-a005-70c0ce68d09c-kube-api-access-p7k6l\") pod \"188ccefd-703d-4857-a005-70c0ce68d09c\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.383325 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/188ccefd-703d-4857-a005-70c0ce68d09c-scripts\") pod \"188ccefd-703d-4857-a005-70c0ce68d09c\" (UID: \"188ccefd-703d-4857-a005-70c0ce68d09c\") " Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.385075 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.385522 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.385541 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcf5b\" (UniqueName: \"kubernetes.io/projected/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-kube-api-access-gcf5b\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.386439 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/188ccefd-703d-4857-a005-70c0ce68d09c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "188ccefd-703d-4857-a005-70c0ce68d09c" (UID: "188ccefd-703d-4857-a005-70c0ce68d09c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.387594 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/188ccefd-703d-4857-a005-70c0ce68d09c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "188ccefd-703d-4857-a005-70c0ce68d09c" (UID: "188ccefd-703d-4857-a005-70c0ce68d09c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.392014 4824 scope.go:117] "RemoveContainer" containerID="fe26ca2e87cd228b8c6a6502631ef82a32112cc0374160767f4acd98b820dd7b" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.435939 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/188ccefd-703d-4857-a005-70c0ce68d09c-scripts" (OuterVolumeSpecName: "scripts") pod "188ccefd-703d-4857-a005-70c0ce68d09c" (UID: "188ccefd-703d-4857-a005-70c0ce68d09c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.437592 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e74eb53-a9de-4679-99d1-7bb097c01d93-kube-api-access-tlg58" (OuterVolumeSpecName: "kube-api-access-tlg58") pod "0e74eb53-a9de-4679-99d1-7bb097c01d93" (UID: "0e74eb53-a9de-4679-99d1-7bb097c01d93"). InnerVolumeSpecName "kube-api-access-tlg58". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.439129 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e74eb53-a9de-4679-99d1-7bb097c01d93-config-data" (OuterVolumeSpecName: "config-data") pod "0e74eb53-a9de-4679-99d1-7bb097c01d93" (UID: "0e74eb53-a9de-4679-99d1-7bb097c01d93"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.471475 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/188ccefd-703d-4857-a005-70c0ce68d09c-kube-api-access-p7k6l" (OuterVolumeSpecName: "kube-api-access-p7k6l") pod "188ccefd-703d-4857-a005-70c0ce68d09c" (UID: "188ccefd-703d-4857-a005-70c0ce68d09c"). InnerVolumeSpecName "kube-api-access-p7k6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.477343 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e74eb53-a9de-4679-99d1-7bb097c01d93-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0e74eb53-a9de-4679-99d1-7bb097c01d93" (UID: "0e74eb53-a9de-4679-99d1-7bb097c01d93"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.484327 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e74eb53-a9de-4679-99d1-7bb097c01d93-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0e74eb53-a9de-4679-99d1-7bb097c01d93" (UID: "0e74eb53-a9de-4679-99d1-7bb097c01d93"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.490407 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/188ccefd-703d-4857-a005-70c0ce68d09c-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.490458 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlg58\" (UniqueName: \"kubernetes.io/projected/0e74eb53-a9de-4679-99d1-7bb097c01d93-kube-api-access-tlg58\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.490473 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7k6l\" (UniqueName: \"kubernetes.io/projected/188ccefd-703d-4857-a005-70c0ce68d09c-kube-api-access-p7k6l\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.490485 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/188ccefd-703d-4857-a005-70c0ce68d09c-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.490496 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e74eb53-a9de-4679-99d1-7bb097c01d93-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.490505 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/188ccefd-703d-4857-a005-70c0ce68d09c-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.490517 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e74eb53-a9de-4679-99d1-7bb097c01d93-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.490527 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e74eb53-a9de-4679-99d1-7bb097c01d93-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.564431 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b" (UID: "6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.586747 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-6fffc89c95-wt6ms" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.592324 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.658245 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/188ccefd-703d-4857-a005-70c0ce68d09c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "188ccefd-703d-4857-a005-70c0ce68d09c" (UID: "188ccefd-703d-4857-a005-70c0ce68d09c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.665723 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-85dcdb56bc-4xfwk"] Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.694285 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/188ccefd-703d-4857-a005-70c0ce68d09c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.694896 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-config-data" (OuterVolumeSpecName: "config-data") pod "6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b" (UID: "6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.796948 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.797562 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/188ccefd-703d-4857-a005-70c0ce68d09c-config-data" (OuterVolumeSpecName: "config-data") pod "188ccefd-703d-4857-a005-70c0ce68d09c" (UID: "188ccefd-703d-4857-a005-70c0ce68d09c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.805306 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/188ccefd-703d-4857-a005-70c0ce68d09c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "188ccefd-703d-4857-a005-70c0ce68d09c" (UID: "188ccefd-703d-4857-a005-70c0ce68d09c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.810859 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7d54cc949d-rzfx4"] Dec 09 10:13:11 crc kubenswrapper[4824]: W1209 10:13:11.820351 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8154a7d_1dbc_4fa3_a862_140f2fd807b2.slice/crio-deaf4df5a37b1d42169f79c4f2bb9cacd88e9626a50040e4e637af0427f8ad8c WatchSource:0}: Error finding container deaf4df5a37b1d42169f79c4f2bb9cacd88e9626a50040e4e637af0427f8ad8c: Status 404 returned error can't find the container with id deaf4df5a37b1d42169f79c4f2bb9cacd88e9626a50040e4e637af0427f8ad8c Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.827145 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-78488f59bc-grx2c"] Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.888442 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-69f49fcbcf-7bnk5"] Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.904986 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188ccefd-703d-4857-a005-70c0ce68d09c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.905037 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188ccefd-703d-4857-a005-70c0ce68d09c-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.952209 4824 scope.go:117] "RemoveContainer" containerID="86c269b7bff445ae59c766a934fd866cc747ab648a33b2f9c4e8ec19aef44a93" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.959294 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-78488f59bc-grx2c" event={"ID":"ddd52d5e-b6bd-403e-affe-c7d25aa4d556","Type":"ContainerStarted","Data":"69ef6324025d05aefef805521350fe9567cdd4d5cdc971040e5dd0da1b09a43f"} Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.959342 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.959359 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.962560 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7d54cc949d-rzfx4" event={"ID":"d8154a7d-1dbc-4fa3-a862-140f2fd807b2","Type":"ContainerStarted","Data":"deaf4df5a37b1d42169f79c4f2bb9cacd88e9626a50040e4e637af0427f8ad8c"} Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.988274 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:13:11 crc kubenswrapper[4824]: I1209 10:13:11.990642 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69f49fcbcf-7bnk5" event={"ID":"ec8287ef-1c22-4524-8a7b-c20d76430002","Type":"ContainerStarted","Data":"5f7c86e281ef96b58b48242f2b73732409e9e17d68ea982e6b47eee4c7c64f50"} Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.002166 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" event={"ID":"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8","Type":"ContainerStarted","Data":"30c1dfbbbf33f5341b9797eb0f33e46c09adec7ea6e20d991ab0e82e9986f5a7"} Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.003436 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.015211 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 09 10:13:12 crc kubenswrapper[4824]: E1209 10:13:12.015572 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="188ccefd-703d-4857-a005-70c0ce68d09c" containerName="proxy-httpd" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.015584 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="188ccefd-703d-4857-a005-70c0ce68d09c" containerName="proxy-httpd" Dec 09 10:13:12 crc kubenswrapper[4824]: E1209 10:13:12.015607 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b" containerName="cinder-api-log" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.015614 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b" containerName="cinder-api-log" Dec 09 10:13:12 crc kubenswrapper[4824]: E1209 10:13:12.015642 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="188ccefd-703d-4857-a005-70c0ce68d09c" containerName="ceilometer-central-agent" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.015649 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="188ccefd-703d-4857-a005-70c0ce68d09c" containerName="ceilometer-central-agent" Dec 09 10:13:12 crc kubenswrapper[4824]: E1209 10:13:12.015673 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="188ccefd-703d-4857-a005-70c0ce68d09c" containerName="ceilometer-notification-agent" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.015681 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="188ccefd-703d-4857-a005-70c0ce68d09c" containerName="ceilometer-notification-agent" Dec 09 10:13:12 crc kubenswrapper[4824]: E1209 10:13:12.015704 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b" containerName="cinder-api" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.015713 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b" containerName="cinder-api" Dec 09 10:13:12 crc kubenswrapper[4824]: E1209 10:13:12.015732 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="188ccefd-703d-4857-a005-70c0ce68d09c" containerName="sg-core" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.015741 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="188ccefd-703d-4857-a005-70c0ce68d09c" containerName="sg-core" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.016476 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="188ccefd-703d-4857-a005-70c0ce68d09c" containerName="ceilometer-notification-agent" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.016492 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="188ccefd-703d-4857-a005-70c0ce68d09c" containerName="proxy-httpd" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.016505 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b" containerName="cinder-api-log" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.016515 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="188ccefd-703d-4857-a005-70c0ce68d09c" containerName="ceilometer-central-agent" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.016527 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="188ccefd-703d-4857-a005-70c0ce68d09c" containerName="sg-core" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.016539 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b" containerName="cinder-api" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.018596 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7466b579b4-qpdj6" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.018603 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-85dcdb56bc-4xfwk" event={"ID":"574c42ce-f75f-4dab-bd77-916b9a4d2c6d","Type":"ContainerStarted","Data":"9a934419701b71e4db4105a37d2dc70de435306ab97a798a39582d21d5a5eaf5"} Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.027922 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.029757 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.047609 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.047866 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.060613 4824 scope.go:117] "RemoveContainer" containerID="09192e79e2b24b4ba4ff045af664c1a16f27e0e694d1b16486716f09b6c84652" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.061203 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.076293 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-79457998d5-fpwv7" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.104664 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.143281 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.161728 4824 scope.go:117] "RemoveContainer" containerID="d4512841b2875155ab8b8e4cd53333f67457e28ae214e9905332d1538133d993" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.165604 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.174393 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:13:12 crc kubenswrapper[4824]: W1209 10:13:12.178913 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6e9dedb_7873_4008_a9a0_fa90ec7c99f1.slice/crio-6d34a1e4bfc2eaeebf01f7268e2d032e85b89e81f2b3d82e9ccb1d7a05f51a28 WatchSource:0}: Error finding container 6d34a1e4bfc2eaeebf01f7268e2d032e85b89e81f2b3d82e9ccb1d7a05f51a28: Status 404 returned error can't find the container with id 6d34a1e4bfc2eaeebf01f7268e2d032e85b89e81f2b3d82e9ccb1d7a05f51a28 Dec 09 10:13:12 crc kubenswrapper[4824]: W1209 10:13:12.181268 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14358ed8_0c10_48f7_a1b1_37e75e7a6400.slice/crio-7c4b1e2e5473b7a828801198d6ff9a886d4e523f0f32395ce793221796eb0132 WatchSource:0}: Error finding container 7c4b1e2e5473b7a828801198d6ff9a886d4e523f0f32395ce793221796eb0132: Status 404 returned error can't find the container with id 7c4b1e2e5473b7a828801198d6ff9a886d4e523f0f32395ce793221796eb0132 Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.181498 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.181591 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.191349 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" podStartSLOduration=22.191325568 podStartE2EDuration="22.191325568s" podCreationTimestamp="2025-12-09 10:12:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:13:12.06163878 +0000 UTC m=+1548.396143477" watchObservedRunningTime="2025-12-09 10:13:12.191325568 +0000 UTC m=+1548.525830235" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.215062 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aeb18289-ca5e-45d5-8976-53d77b8c8ee9-combined-ca-bundle\") pod \"aeb18289-ca5e-45d5-8976-53d77b8c8ee9\" (UID: \"aeb18289-ca5e-45d5-8976-53d77b8c8ee9\") " Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.215227 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aeb18289-ca5e-45d5-8976-53d77b8c8ee9-config-data-custom\") pod \"aeb18289-ca5e-45d5-8976-53d77b8c8ee9\" (UID: \"aeb18289-ca5e-45d5-8976-53d77b8c8ee9\") " Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.215294 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdlzz\" (UniqueName: \"kubernetes.io/projected/aeb18289-ca5e-45d5-8976-53d77b8c8ee9-kube-api-access-vdlzz\") pod \"aeb18289-ca5e-45d5-8976-53d77b8c8ee9\" (UID: \"aeb18289-ca5e-45d5-8976-53d77b8c8ee9\") " Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.215323 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aeb18289-ca5e-45d5-8976-53d77b8c8ee9-config-data\") pod \"aeb18289-ca5e-45d5-8976-53d77b8c8ee9\" (UID: \"aeb18289-ca5e-45d5-8976-53d77b8c8ee9\") " Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.216317 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0755044-204f-4e63-a3cb-69b7c8b64159-config-data\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.216355 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0755044-204f-4e63-a3cb-69b7c8b64159-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.216389 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggbw4\" (UniqueName: \"kubernetes.io/projected/9cc85a9e-8935-432d-8324-491c402402ea-kube-api-access-ggbw4\") pod \"ceilometer-0\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " pod="openstack/ceilometer-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.216433 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0755044-204f-4e63-a3cb-69b7c8b64159-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.216455 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cc85a9e-8935-432d-8324-491c402402ea-config-data\") pod \"ceilometer-0\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " pod="openstack/ceilometer-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.216469 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0755044-204f-4e63-a3cb-69b7c8b64159-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.216487 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0755044-204f-4e63-a3cb-69b7c8b64159-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.216508 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cc85a9e-8935-432d-8324-491c402402ea-scripts\") pod \"ceilometer-0\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " pod="openstack/ceilometer-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.216537 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cc85a9e-8935-432d-8324-491c402402ea-log-httpd\") pod \"ceilometer-0\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " pod="openstack/ceilometer-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.216578 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0755044-204f-4e63-a3cb-69b7c8b64159-logs\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.216676 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0755044-204f-4e63-a3cb-69b7c8b64159-config-data-custom\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.216709 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cc85a9e-8935-432d-8324-491c402402ea-run-httpd\") pod \"ceilometer-0\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " pod="openstack/ceilometer-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.216758 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggwnf\" (UniqueName: \"kubernetes.io/projected/e0755044-204f-4e63-a3cb-69b7c8b64159-kube-api-access-ggwnf\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.216812 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9cc85a9e-8935-432d-8324-491c402402ea-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " pod="openstack/ceilometer-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.216827 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cc85a9e-8935-432d-8324-491c402402ea-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " pod="openstack/ceilometer-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.216848 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0755044-204f-4e63-a3cb-69b7c8b64159-scripts\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.220019 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aeb18289-ca5e-45d5-8976-53d77b8c8ee9-kube-api-access-vdlzz" (OuterVolumeSpecName: "kube-api-access-vdlzz") pod "aeb18289-ca5e-45d5-8976-53d77b8c8ee9" (UID: "aeb18289-ca5e-45d5-8976-53d77b8c8ee9"). InnerVolumeSpecName "kube-api-access-vdlzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.222155 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeb18289-ca5e-45d5-8976-53d77b8c8ee9-config-data" (OuterVolumeSpecName: "config-data") pod "aeb18289-ca5e-45d5-8976-53d77b8c8ee9" (UID: "aeb18289-ca5e-45d5-8976-53d77b8c8ee9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.226141 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeb18289-ca5e-45d5-8976-53d77b8c8ee9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "aeb18289-ca5e-45d5-8976-53d77b8c8ee9" (UID: "aeb18289-ca5e-45d5-8976-53d77b8c8ee9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.226880 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.251141 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeb18289-ca5e-45d5-8976-53d77b8c8ee9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aeb18289-ca5e-45d5-8976-53d77b8c8ee9" (UID: "aeb18289-ca5e-45d5-8976-53d77b8c8ee9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.262633 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-98f88b69-9xf7z"] Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.292304 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-ff546578f-d6scd"] Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.312206 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7466b579b4-qpdj6"] Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.321835 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggbw4\" (UniqueName: \"kubernetes.io/projected/9cc85a9e-8935-432d-8324-491c402402ea-kube-api-access-ggbw4\") pod \"ceilometer-0\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " pod="openstack/ceilometer-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.321977 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0755044-204f-4e63-a3cb-69b7c8b64159-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.322158 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cc85a9e-8935-432d-8324-491c402402ea-config-data\") pod \"ceilometer-0\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " pod="openstack/ceilometer-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.322190 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0755044-204f-4e63-a3cb-69b7c8b64159-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.322261 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0755044-204f-4e63-a3cb-69b7c8b64159-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.323725 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cc85a9e-8935-432d-8324-491c402402ea-scripts\") pod \"ceilometer-0\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " pod="openstack/ceilometer-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.322366 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0755044-204f-4e63-a3cb-69b7c8b64159-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.323810 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cc85a9e-8935-432d-8324-491c402402ea-log-httpd\") pod \"ceilometer-0\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " pod="openstack/ceilometer-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.324308 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0755044-204f-4e63-a3cb-69b7c8b64159-logs\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.324472 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cc85a9e-8935-432d-8324-491c402402ea-log-httpd\") pod \"ceilometer-0\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " pod="openstack/ceilometer-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.324574 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0755044-204f-4e63-a3cb-69b7c8b64159-config-data-custom\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.324636 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cc85a9e-8935-432d-8324-491c402402ea-run-httpd\") pod \"ceilometer-0\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " pod="openstack/ceilometer-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.324733 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggwnf\" (UniqueName: \"kubernetes.io/projected/e0755044-204f-4e63-a3cb-69b7c8b64159-kube-api-access-ggwnf\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.324847 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9cc85a9e-8935-432d-8324-491c402402ea-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " pod="openstack/ceilometer-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.324874 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cc85a9e-8935-432d-8324-491c402402ea-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " pod="openstack/ceilometer-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.324917 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0755044-204f-4e63-a3cb-69b7c8b64159-scripts\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.324952 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0755044-204f-4e63-a3cb-69b7c8b64159-logs\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.325150 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0755044-204f-4e63-a3cb-69b7c8b64159-config-data\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.325184 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0755044-204f-4e63-a3cb-69b7c8b64159-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.327928 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aeb18289-ca5e-45d5-8976-53d77b8c8ee9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.328395 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cc85a9e-8935-432d-8324-491c402402ea-run-httpd\") pod \"ceilometer-0\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " pod="openstack/ceilometer-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.329104 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aeb18289-ca5e-45d5-8976-53d77b8c8ee9-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.329142 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdlzz\" (UniqueName: \"kubernetes.io/projected/aeb18289-ca5e-45d5-8976-53d77b8c8ee9-kube-api-access-vdlzz\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.329162 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aeb18289-ca5e-45d5-8976-53d77b8c8ee9-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.337442 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cc85a9e-8935-432d-8324-491c402402ea-scripts\") pod \"ceilometer-0\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " pod="openstack/ceilometer-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.338804 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0755044-204f-4e63-a3cb-69b7c8b64159-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.349102 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-7466b579b4-qpdj6"] Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.354292 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0755044-204f-4e63-a3cb-69b7c8b64159-config-data\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.357887 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggbw4\" (UniqueName: \"kubernetes.io/projected/9cc85a9e-8935-432d-8324-491c402402ea-kube-api-access-ggbw4\") pod \"ceilometer-0\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " pod="openstack/ceilometer-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.357973 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0755044-204f-4e63-a3cb-69b7c8b64159-config-data-custom\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.357996 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cc85a9e-8935-432d-8324-491c402402ea-config-data\") pod \"ceilometer-0\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " pod="openstack/ceilometer-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.358701 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cc85a9e-8935-432d-8324-491c402402ea-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " pod="openstack/ceilometer-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.360245 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9cc85a9e-8935-432d-8324-491c402402ea-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " pod="openstack/ceilometer-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.360520 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0755044-204f-4e63-a3cb-69b7c8b64159-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.361070 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0755044-204f-4e63-a3cb-69b7c8b64159-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.362824 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggwnf\" (UniqueName: \"kubernetes.io/projected/e0755044-204f-4e63-a3cb-69b7c8b64159-kube-api-access-ggwnf\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.363291 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0755044-204f-4e63-a3cb-69b7c8b64159-scripts\") pod \"cinder-api-0\" (UID: \"e0755044-204f-4e63-a3cb-69b7c8b64159\") " pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.430052 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 10:13:12 crc kubenswrapper[4824]: I1209 10:13:12.525408 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.055624 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-79457998d5-fpwv7" event={"ID":"aeb18289-ca5e-45d5-8976-53d77b8c8ee9","Type":"ContainerDied","Data":"fd16e607756ce62917f50427e7731ced31e470aa377d84ebc1c0b839ac791f45"} Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.056565 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-79457998d5-fpwv7" Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.062242 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-98f88b69-9xf7z" event={"ID":"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1","Type":"ContainerStarted","Data":"6d34a1e4bfc2eaeebf01f7268e2d032e85b89e81f2b3d82e9ccb1d7a05f51a28"} Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.091560 4824 scope.go:117] "RemoveContainer" containerID="22653aa1edc638e086ee4259cc176ab30f46152ff9de8640030b571df4251c42" Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.091660 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-78488f59bc-grx2c" event={"ID":"ddd52d5e-b6bd-403e-affe-c7d25aa4d556","Type":"ContainerStarted","Data":"22653aa1edc638e086ee4259cc176ab30f46152ff9de8640030b571df4251c42"} Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.101382 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-ff546578f-d6scd" event={"ID":"14358ed8-0c10-48f7-a1b1-37e75e7a6400","Type":"ContainerStarted","Data":"7c4b1e2e5473b7a828801198d6ff9a886d4e523f0f32395ce793221796eb0132"} Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.104967 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69f49fcbcf-7bnk5" event={"ID":"ec8287ef-1c22-4524-8a7b-c20d76430002","Type":"ContainerStarted","Data":"cf0b129f3db23503ff45cd7691ce785e15127b9840e9c28cc0f27060e0efb841"} Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.105219 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-69f49fcbcf-7bnk5" Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.106917 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.147384 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-85dcdb56bc-4xfwk" event={"ID":"574c42ce-f75f-4dab-bd77-916b9a4d2c6d","Type":"ContainerStarted","Data":"1253ea5cce9374b6db7702d0f71e2af07fea239c958fc319fedc5494666bd071"} Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.147440 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-85dcdb56bc-4xfwk" event={"ID":"574c42ce-f75f-4dab-bd77-916b9a4d2c6d","Type":"ContainerStarted","Data":"bfe77362615d7dce1ef8562bd07a6657d8a78847090ebd903382b7d075612143"} Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.148617 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.148658 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.186120 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-69f49fcbcf-7bnk5" podStartSLOduration=11.590338368 podStartE2EDuration="12.18609393s" podCreationTimestamp="2025-12-09 10:13:01 +0000 UTC" firstStartedPulling="2025-12-09 10:13:11.950512263 +0000 UTC m=+1548.285016930" lastFinishedPulling="2025-12-09 10:13:12.546267825 +0000 UTC m=+1548.880772492" observedRunningTime="2025-12-09 10:13:13.132731643 +0000 UTC m=+1549.467236310" watchObservedRunningTime="2025-12-09 10:13:13.18609393 +0000 UTC m=+1549.520598597" Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.197094 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7d54cc949d-rzfx4" event={"ID":"d8154a7d-1dbc-4fa3-a862-140f2fd807b2","Type":"ContainerStarted","Data":"6a711400741bf3c06a5572c6a46ac91b8ec31fb5fdfda32e199126f0b080b249"} Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.198052 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-7d54cc949d-rzfx4" Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.215193 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-79457998d5-fpwv7"] Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.241124 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-79457998d5-fpwv7"] Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.244429 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-85dcdb56bc-4xfwk" podStartSLOduration=11.244405117 podStartE2EDuration="11.244405117s" podCreationTimestamp="2025-12-09 10:13:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:13:13.216706736 +0000 UTC m=+1549.551211413" watchObservedRunningTime="2025-12-09 10:13:13.244405117 +0000 UTC m=+1549.578909784" Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.263316 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.273494 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-7d54cc949d-rzfx4" podStartSLOduration=13.273461151 podStartE2EDuration="13.273461151s" podCreationTimestamp="2025-12-09 10:13:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:13:13.241093002 +0000 UTC m=+1549.575597669" watchObservedRunningTime="2025-12-09 10:13:13.273461151 +0000 UTC m=+1549.607965828" Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.948207 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e74eb53-a9de-4679-99d1-7bb097c01d93" path="/var/lib/kubelet/pods/0e74eb53-a9de-4679-99d1-7bb097c01d93/volumes" Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.949058 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="188ccefd-703d-4857-a005-70c0ce68d09c" path="/var/lib/kubelet/pods/188ccefd-703d-4857-a005-70c0ce68d09c/volumes" Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.959427 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b" path="/var/lib/kubelet/pods/6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b/volumes" Dec 09 10:13:13 crc kubenswrapper[4824]: I1209 10:13:13.960848 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aeb18289-ca5e-45d5-8976-53d77b8c8ee9" path="/var/lib/kubelet/pods/aeb18289-ca5e-45d5-8976-53d77b8c8ee9/volumes" Dec 09 10:13:14 crc kubenswrapper[4824]: I1209 10:13:14.122813 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="6f0f6c25-ef0d-418f-a1ea-eccfe9d0613b" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.205:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 10:13:14 crc kubenswrapper[4824]: I1209 10:13:14.220159 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-ff546578f-d6scd" event={"ID":"14358ed8-0c10-48f7-a1b1-37e75e7a6400","Type":"ContainerStarted","Data":"4d69b95b0579a8d3ad4ea33d1c22f7fdffb90d75e0a488beff69658977be69de"} Dec 09 10:13:14 crc kubenswrapper[4824]: I1209 10:13:14.220452 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-ff546578f-d6scd" Dec 09 10:13:14 crc kubenswrapper[4824]: I1209 10:13:14.225049 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e0755044-204f-4e63-a3cb-69b7c8b64159","Type":"ContainerStarted","Data":"1afacbcd4df9fa583c623409b38bd25495f16305a93cb41a7c9e4204b9676112"} Dec 09 10:13:14 crc kubenswrapper[4824]: I1209 10:13:14.227935 4824 generic.go:334] "Generic (PLEG): container finished" podID="ec8287ef-1c22-4524-8a7b-c20d76430002" containerID="cf0b129f3db23503ff45cd7691ce785e15127b9840e9c28cc0f27060e0efb841" exitCode=1 Dec 09 10:13:14 crc kubenswrapper[4824]: I1209 10:13:14.228101 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69f49fcbcf-7bnk5" event={"ID":"ec8287ef-1c22-4524-8a7b-c20d76430002","Type":"ContainerDied","Data":"cf0b129f3db23503ff45cd7691ce785e15127b9840e9c28cc0f27060e0efb841"} Dec 09 10:13:14 crc kubenswrapper[4824]: I1209 10:13:14.229398 4824 scope.go:117] "RemoveContainer" containerID="cf0b129f3db23503ff45cd7691ce785e15127b9840e9c28cc0f27060e0efb841" Dec 09 10:13:14 crc kubenswrapper[4824]: I1209 10:13:14.232434 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-98f88b69-9xf7z" event={"ID":"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1","Type":"ContainerStarted","Data":"1004fb0fd749d85593aaa49a6c0f575039b7bf13fb8560d1104508bdc20a30fe"} Dec 09 10:13:14 crc kubenswrapper[4824]: I1209 10:13:14.232813 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-98f88b69-9xf7z" Dec 09 10:13:14 crc kubenswrapper[4824]: I1209 10:13:14.234160 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cc85a9e-8935-432d-8324-491c402402ea","Type":"ContainerStarted","Data":"5564e132583313a8fd5642f07f74cc3424deeb1d6809b582acbcc982bd50ab75"} Dec 09 10:13:14 crc kubenswrapper[4824]: I1209 10:13:14.251344 4824 generic.go:334] "Generic (PLEG): container finished" podID="ddd52d5e-b6bd-403e-affe-c7d25aa4d556" containerID="22653aa1edc638e086ee4259cc176ab30f46152ff9de8640030b571df4251c42" exitCode=1 Dec 09 10:13:14 crc kubenswrapper[4824]: I1209 10:13:14.251861 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-78488f59bc-grx2c" event={"ID":"ddd52d5e-b6bd-403e-affe-c7d25aa4d556","Type":"ContainerDied","Data":"22653aa1edc638e086ee4259cc176ab30f46152ff9de8640030b571df4251c42"} Dec 09 10:13:14 crc kubenswrapper[4824]: I1209 10:13:14.251937 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-78488f59bc-grx2c" event={"ID":"ddd52d5e-b6bd-403e-affe-c7d25aa4d556","Type":"ContainerStarted","Data":"f690aba191f16dc81b5dcbfdedec49c1b3a9ddda35231e2e292c4e92691db354"} Dec 09 10:13:14 crc kubenswrapper[4824]: I1209 10:13:14.252885 4824 status_manager.go:317] "Container readiness changed for unknown container" pod="openstack/heat-cfnapi-78488f59bc-grx2c" containerID="cri-o://22653aa1edc638e086ee4259cc176ab30f46152ff9de8640030b571df4251c42" Dec 09 10:13:14 crc kubenswrapper[4824]: I1209 10:13:14.252935 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-78488f59bc-grx2c" Dec 09 10:13:14 crc kubenswrapper[4824]: I1209 10:13:14.282025 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-ff546578f-d6scd" podStartSLOduration=9.747311204 podStartE2EDuration="10.282002103s" podCreationTimestamp="2025-12-09 10:13:04 +0000 UTC" firstStartedPulling="2025-12-09 10:13:12.18761991 +0000 UTC m=+1548.522124577" lastFinishedPulling="2025-12-09 10:13:12.722310809 +0000 UTC m=+1549.056815476" observedRunningTime="2025-12-09 10:13:14.242618389 +0000 UTC m=+1550.577123066" watchObservedRunningTime="2025-12-09 10:13:14.282002103 +0000 UTC m=+1550.616506770" Dec 09 10:13:14 crc kubenswrapper[4824]: I1209 10:13:14.315139 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-98f88b69-9xf7z" podStartSLOduration=9.825817083 podStartE2EDuration="10.315099416s" podCreationTimestamp="2025-12-09 10:13:04 +0000 UTC" firstStartedPulling="2025-12-09 10:13:12.187901139 +0000 UTC m=+1548.522405806" lastFinishedPulling="2025-12-09 10:13:12.677183472 +0000 UTC m=+1549.011688139" observedRunningTime="2025-12-09 10:13:14.268485483 +0000 UTC m=+1550.602990150" watchObservedRunningTime="2025-12-09 10:13:14.315099416 +0000 UTC m=+1550.649604103" Dec 09 10:13:14 crc kubenswrapper[4824]: I1209 10:13:14.354194 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-78488f59bc-grx2c" podStartSLOduration=12.809668228 podStartE2EDuration="13.35416675s" podCreationTimestamp="2025-12-09 10:13:01 +0000 UTC" firstStartedPulling="2025-12-09 10:13:11.829375057 +0000 UTC m=+1548.163879724" lastFinishedPulling="2025-12-09 10:13:12.373873579 +0000 UTC m=+1548.708378246" observedRunningTime="2025-12-09 10:13:14.311357998 +0000 UTC m=+1550.645862675" watchObservedRunningTime="2025-12-09 10:13:14.35416675 +0000 UTC m=+1550.688671417" Dec 09 10:13:15 crc kubenswrapper[4824]: I1209 10:13:15.271523 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cc85a9e-8935-432d-8324-491c402402ea","Type":"ContainerStarted","Data":"26dbed783b66c0024fba06c893071ac5841a36aa5722a623cd78e75bc9f4725b"} Dec 09 10:13:15 crc kubenswrapper[4824]: I1209 10:13:15.279908 4824 generic.go:334] "Generic (PLEG): container finished" podID="ddd52d5e-b6bd-403e-affe-c7d25aa4d556" containerID="f690aba191f16dc81b5dcbfdedec49c1b3a9ddda35231e2e292c4e92691db354" exitCode=1 Dec 09 10:13:15 crc kubenswrapper[4824]: I1209 10:13:15.279991 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-78488f59bc-grx2c" event={"ID":"ddd52d5e-b6bd-403e-affe-c7d25aa4d556","Type":"ContainerDied","Data":"f690aba191f16dc81b5dcbfdedec49c1b3a9ddda35231e2e292c4e92691db354"} Dec 09 10:13:15 crc kubenswrapper[4824]: I1209 10:13:15.280098 4824 scope.go:117] "RemoveContainer" containerID="22653aa1edc638e086ee4259cc176ab30f46152ff9de8640030b571df4251c42" Dec 09 10:13:15 crc kubenswrapper[4824]: I1209 10:13:15.280758 4824 scope.go:117] "RemoveContainer" containerID="f690aba191f16dc81b5dcbfdedec49c1b3a9ddda35231e2e292c4e92691db354" Dec 09 10:13:15 crc kubenswrapper[4824]: E1209 10:13:15.281074 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-78488f59bc-grx2c_openstack(ddd52d5e-b6bd-403e-affe-c7d25aa4d556)\"" pod="openstack/heat-cfnapi-78488f59bc-grx2c" podUID="ddd52d5e-b6bd-403e-affe-c7d25aa4d556" Dec 09 10:13:15 crc kubenswrapper[4824]: I1209 10:13:15.289475 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e0755044-204f-4e63-a3cb-69b7c8b64159","Type":"ContainerStarted","Data":"931a364374dc2babc2eb5bdbfac51f3ef7f45ee9b609c7898d4f4bb3df824058"} Dec 09 10:13:16 crc kubenswrapper[4824]: I1209 10:13:16.299951 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" Dec 09 10:13:16 crc kubenswrapper[4824]: I1209 10:13:16.305410 4824 scope.go:117] "RemoveContainer" containerID="f690aba191f16dc81b5dcbfdedec49c1b3a9ddda35231e2e292c4e92691db354" Dec 09 10:13:16 crc kubenswrapper[4824]: E1209 10:13:16.305688 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-78488f59bc-grx2c_openstack(ddd52d5e-b6bd-403e-affe-c7d25aa4d556)\"" pod="openstack/heat-cfnapi-78488f59bc-grx2c" podUID="ddd52d5e-b6bd-403e-affe-c7d25aa4d556" Dec 09 10:13:16 crc kubenswrapper[4824]: I1209 10:13:16.386387 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-jgv7h"] Dec 09 10:13:16 crc kubenswrapper[4824]: I1209 10:13:16.386699 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" podUID="6b816dc5-829b-4c38-b2e0-b08cfbfc1617" containerName="dnsmasq-dns" containerID="cri-o://29c9a683ae0d4909abd63d7f40848eeabda0f73ae8dea7e1587f0586061f08a4" gracePeriod=10 Dec 09 10:13:17 crc kubenswrapper[4824]: I1209 10:13:17.039981 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-78488f59bc-grx2c" Dec 09 10:13:17 crc kubenswrapper[4824]: I1209 10:13:17.040524 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-78488f59bc-grx2c" Dec 09 10:13:17 crc kubenswrapper[4824]: I1209 10:13:17.057937 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-69f49fcbcf-7bnk5" Dec 09 10:13:17 crc kubenswrapper[4824]: I1209 10:13:17.333456 4824 scope.go:117] "RemoveContainer" containerID="f690aba191f16dc81b5dcbfdedec49c1b3a9ddda35231e2e292c4e92691db354" Dec 09 10:13:17 crc kubenswrapper[4824]: E1209 10:13:17.334204 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-78488f59bc-grx2c_openstack(ddd52d5e-b6bd-403e-affe-c7d25aa4d556)\"" pod="openstack/heat-cfnapi-78488f59bc-grx2c" podUID="ddd52d5e-b6bd-403e-affe-c7d25aa4d556" Dec 09 10:13:17 crc kubenswrapper[4824]: I1209 10:13:17.337183 4824 generic.go:334] "Generic (PLEG): container finished" podID="6b816dc5-829b-4c38-b2e0-b08cfbfc1617" containerID="29c9a683ae0d4909abd63d7f40848eeabda0f73ae8dea7e1587f0586061f08a4" exitCode=0 Dec 09 10:13:17 crc kubenswrapper[4824]: I1209 10:13:17.337271 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" event={"ID":"6b816dc5-829b-4c38-b2e0-b08cfbfc1617","Type":"ContainerDied","Data":"29c9a683ae0d4909abd63d7f40848eeabda0f73ae8dea7e1587f0586061f08a4"} Dec 09 10:13:17 crc kubenswrapper[4824]: I1209 10:13:17.340380 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e0755044-204f-4e63-a3cb-69b7c8b64159","Type":"ContainerStarted","Data":"c23e56b97f4eab673b67a99c488dfa360d7992ceac36277b5d7d197395edfac8"} Dec 09 10:13:17 crc kubenswrapper[4824]: I1209 10:13:17.340944 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 09 10:13:17 crc kubenswrapper[4824]: I1209 10:13:17.358940 4824 generic.go:334] "Generic (PLEG): container finished" podID="ec8287ef-1c22-4524-8a7b-c20d76430002" containerID="466d0a1c1b79a9bf590cfd803ef7a2afed5130e42ae2e0f32a8e3b35a4037bbf" exitCode=1 Dec 09 10:13:17 crc kubenswrapper[4824]: I1209 10:13:17.359061 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69f49fcbcf-7bnk5" event={"ID":"ec8287ef-1c22-4524-8a7b-c20d76430002","Type":"ContainerDied","Data":"466d0a1c1b79a9bf590cfd803ef7a2afed5130e42ae2e0f32a8e3b35a4037bbf"} Dec 09 10:13:17 crc kubenswrapper[4824]: I1209 10:13:17.359099 4824 scope.go:117] "RemoveContainer" containerID="cf0b129f3db23503ff45cd7691ce785e15127b9840e9c28cc0f27060e0efb841" Dec 09 10:13:17 crc kubenswrapper[4824]: I1209 10:13:17.359948 4824 scope.go:117] "RemoveContainer" containerID="466d0a1c1b79a9bf590cfd803ef7a2afed5130e42ae2e0f32a8e3b35a4037bbf" Dec 09 10:13:17 crc kubenswrapper[4824]: E1209 10:13:17.360289 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-69f49fcbcf-7bnk5_openstack(ec8287ef-1c22-4524-8a7b-c20d76430002)\"" pod="openstack/heat-api-69f49fcbcf-7bnk5" podUID="ec8287ef-1c22-4524-8a7b-c20d76430002" Dec 09 10:13:17 crc kubenswrapper[4824]: I1209 10:13:17.368148 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.368125523 podStartE2EDuration="6.368125523s" podCreationTimestamp="2025-12-09 10:13:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:13:17.363260657 +0000 UTC m=+1553.697765324" watchObservedRunningTime="2025-12-09 10:13:17.368125523 +0000 UTC m=+1553.702630190" Dec 09 10:13:17 crc kubenswrapper[4824]: I1209 10:13:17.438025 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cc85a9e-8935-432d-8324-491c402402ea","Type":"ContainerStarted","Data":"6de41f253743baba6c5372b71ee72fd1c62f8cdb896ddea6e5984165758cf820"} Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.023354 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.121284 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-dns-svc\") pod \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\" (UID: \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\") " Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.121952 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-ovsdbserver-sb\") pod \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\" (UID: \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\") " Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.122023 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cggz\" (UniqueName: \"kubernetes.io/projected/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-kube-api-access-7cggz\") pod \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\" (UID: \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\") " Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.122082 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-dns-swift-storage-0\") pod \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\" (UID: \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\") " Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.122739 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-ovsdbserver-nb\") pod \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\" (UID: \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\") " Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.122772 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-config\") pod \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\" (UID: \"6b816dc5-829b-4c38-b2e0-b08cfbfc1617\") " Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.140603 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-kube-api-access-7cggz" (OuterVolumeSpecName: "kube-api-access-7cggz") pod "6b816dc5-829b-4c38-b2e0-b08cfbfc1617" (UID: "6b816dc5-829b-4c38-b2e0-b08cfbfc1617"). InnerVolumeSpecName "kube-api-access-7cggz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.208565 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.210127 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-85dcdb56bc-4xfwk" Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.225900 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6b816dc5-829b-4c38-b2e0-b08cfbfc1617" (UID: "6b816dc5-829b-4c38-b2e0-b08cfbfc1617"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.227727 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6b816dc5-829b-4c38-b2e0-b08cfbfc1617" (UID: "6b816dc5-829b-4c38-b2e0-b08cfbfc1617"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.232100 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cggz\" (UniqueName: \"kubernetes.io/projected/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-kube-api-access-7cggz\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.232341 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.232408 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.242461 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6b816dc5-829b-4c38-b2e0-b08cfbfc1617" (UID: "6b816dc5-829b-4c38-b2e0-b08cfbfc1617"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.250704 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-config" (OuterVolumeSpecName: "config") pod "6b816dc5-829b-4c38-b2e0-b08cfbfc1617" (UID: "6b816dc5-829b-4c38-b2e0-b08cfbfc1617"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.264879 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6b816dc5-829b-4c38-b2e0-b08cfbfc1617" (UID: "6b816dc5-829b-4c38-b2e0-b08cfbfc1617"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.342358 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.342407 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.342424 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b816dc5-829b-4c38-b2e0-b08cfbfc1617-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.472174 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cc85a9e-8935-432d-8324-491c402402ea","Type":"ContainerStarted","Data":"3f141fcc82254eac5845a2bab2c48e08301ff3b94134665db9e95e6607142bb2"} Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.477036 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" event={"ID":"6b816dc5-829b-4c38-b2e0-b08cfbfc1617","Type":"ContainerDied","Data":"85a0480dda0da6396564457bad6265b5bbf437b535fb5c1209c8ab3781aff40a"} Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.477092 4824 scope.go:117] "RemoveContainer" containerID="29c9a683ae0d4909abd63d7f40848eeabda0f73ae8dea7e1587f0586061f08a4" Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.477268 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-jgv7h" Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.486527 4824 scope.go:117] "RemoveContainer" containerID="466d0a1c1b79a9bf590cfd803ef7a2afed5130e42ae2e0f32a8e3b35a4037bbf" Dec 09 10:13:18 crc kubenswrapper[4824]: E1209 10:13:18.486722 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-69f49fcbcf-7bnk5_openstack(ec8287ef-1c22-4524-8a7b-c20d76430002)\"" pod="openstack/heat-api-69f49fcbcf-7bnk5" podUID="ec8287ef-1c22-4524-8a7b-c20d76430002" Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.487330 4824 scope.go:117] "RemoveContainer" containerID="f690aba191f16dc81b5dcbfdedec49c1b3a9ddda35231e2e292c4e92691db354" Dec 09 10:13:18 crc kubenswrapper[4824]: E1209 10:13:18.487508 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-78488f59bc-grx2c_openstack(ddd52d5e-b6bd-403e-affe-c7d25aa4d556)\"" pod="openstack/heat-cfnapi-78488f59bc-grx2c" podUID="ddd52d5e-b6bd-403e-affe-c7d25aa4d556" Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.509847 4824 scope.go:117] "RemoveContainer" containerID="114b030ac8493b55488b9a7a828053a7e865af0e63128f8d2ffb9d91a3d972f7" Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.587850 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-jgv7h"] Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.599494 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-jgv7h"] Dec 09 10:13:18 crc kubenswrapper[4824]: I1209 10:13:18.612222 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:13:19 crc kubenswrapper[4824]: I1209 10:13:19.926159 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b816dc5-829b-4c38-b2e0-b08cfbfc1617" path="/var/lib/kubelet/pods/6b816dc5-829b-4c38-b2e0-b08cfbfc1617/volumes" Dec 09 10:13:20 crc kubenswrapper[4824]: I1209 10:13:20.520165 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cc85a9e-8935-432d-8324-491c402402ea","Type":"ContainerStarted","Data":"70c869b7ae088e7888ff3f6ef92e2c78807bd104485f5bafd6f17a9276483746"} Dec 09 10:13:20 crc kubenswrapper[4824]: I1209 10:13:20.520525 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 10:13:20 crc kubenswrapper[4824]: I1209 10:13:20.520363 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9cc85a9e-8935-432d-8324-491c402402ea" containerName="ceilometer-notification-agent" containerID="cri-o://6de41f253743baba6c5372b71ee72fd1c62f8cdb896ddea6e5984165758cf820" gracePeriod=30 Dec 09 10:13:20 crc kubenswrapper[4824]: I1209 10:13:20.520326 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9cc85a9e-8935-432d-8324-491c402402ea" containerName="proxy-httpd" containerID="cri-o://70c869b7ae088e7888ff3f6ef92e2c78807bd104485f5bafd6f17a9276483746" gracePeriod=30 Dec 09 10:13:20 crc kubenswrapper[4824]: I1209 10:13:20.520399 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9cc85a9e-8935-432d-8324-491c402402ea" containerName="sg-core" containerID="cri-o://3f141fcc82254eac5845a2bab2c48e08301ff3b94134665db9e95e6607142bb2" gracePeriod=30 Dec 09 10:13:20 crc kubenswrapper[4824]: I1209 10:13:20.520394 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9cc85a9e-8935-432d-8324-491c402402ea" containerName="ceilometer-central-agent" containerID="cri-o://26dbed783b66c0024fba06c893071ac5841a36aa5722a623cd78e75bc9f4725b" gracePeriod=30 Dec 09 10:13:20 crc kubenswrapper[4824]: I1209 10:13:20.546112 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.288370296 podStartE2EDuration="8.546079784s" podCreationTimestamp="2025-12-09 10:13:12 +0000 UTC" firstStartedPulling="2025-12-09 10:13:13.297428285 +0000 UTC m=+1549.631932952" lastFinishedPulling="2025-12-09 10:13:19.555137773 +0000 UTC m=+1555.889642440" observedRunningTime="2025-12-09 10:13:20.540983852 +0000 UTC m=+1556.875488529" watchObservedRunningTime="2025-12-09 10:13:20.546079784 +0000 UTC m=+1556.880584451" Dec 09 10:13:21 crc kubenswrapper[4824]: I1209 10:13:21.538165 4824 generic.go:334] "Generic (PLEG): container finished" podID="9cc85a9e-8935-432d-8324-491c402402ea" containerID="70c869b7ae088e7888ff3f6ef92e2c78807bd104485f5bafd6f17a9276483746" exitCode=0 Dec 09 10:13:21 crc kubenswrapper[4824]: I1209 10:13:21.539369 4824 generic.go:334] "Generic (PLEG): container finished" podID="9cc85a9e-8935-432d-8324-491c402402ea" containerID="3f141fcc82254eac5845a2bab2c48e08301ff3b94134665db9e95e6607142bb2" exitCode=2 Dec 09 10:13:21 crc kubenswrapper[4824]: I1209 10:13:21.539467 4824 generic.go:334] "Generic (PLEG): container finished" podID="9cc85a9e-8935-432d-8324-491c402402ea" containerID="6de41f253743baba6c5372b71ee72fd1c62f8cdb896ddea6e5984165758cf820" exitCode=0 Dec 09 10:13:21 crc kubenswrapper[4824]: I1209 10:13:21.538243 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cc85a9e-8935-432d-8324-491c402402ea","Type":"ContainerDied","Data":"70c869b7ae088e7888ff3f6ef92e2c78807bd104485f5bafd6f17a9276483746"} Dec 09 10:13:21 crc kubenswrapper[4824]: I1209 10:13:21.539611 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cc85a9e-8935-432d-8324-491c402402ea","Type":"ContainerDied","Data":"3f141fcc82254eac5845a2bab2c48e08301ff3b94134665db9e95e6607142bb2"} Dec 09 10:13:21 crc kubenswrapper[4824]: I1209 10:13:21.539652 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cc85a9e-8935-432d-8324-491c402402ea","Type":"ContainerDied","Data":"6de41f253743baba6c5372b71ee72fd1c62f8cdb896ddea6e5984165758cf820"} Dec 09 10:13:22 crc kubenswrapper[4824]: I1209 10:13:22.058114 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-69f49fcbcf-7bnk5" Dec 09 10:13:22 crc kubenswrapper[4824]: I1209 10:13:22.059005 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-69f49fcbcf-7bnk5" Dec 09 10:13:22 crc kubenswrapper[4824]: I1209 10:13:22.059448 4824 scope.go:117] "RemoveContainer" containerID="466d0a1c1b79a9bf590cfd803ef7a2afed5130e42ae2e0f32a8e3b35a4037bbf" Dec 09 10:13:22 crc kubenswrapper[4824]: E1209 10:13:22.059950 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-69f49fcbcf-7bnk5_openstack(ec8287ef-1c22-4524-8a7b-c20d76430002)\"" pod="openstack/heat-api-69f49fcbcf-7bnk5" podUID="ec8287ef-1c22-4524-8a7b-c20d76430002" Dec 09 10:13:22 crc kubenswrapper[4824]: I1209 10:13:22.079951 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-98f88b69-9xf7z" Dec 09 10:13:22 crc kubenswrapper[4824]: I1209 10:13:22.141190 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-ff546578f-d6scd" Dec 09 10:13:22 crc kubenswrapper[4824]: I1209 10:13:22.168206 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-69f49fcbcf-7bnk5"] Dec 09 10:13:22 crc kubenswrapper[4824]: I1209 10:13:22.231569 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-78488f59bc-grx2c"] Dec 09 10:13:22 crc kubenswrapper[4824]: I1209 10:13:22.564102 4824 scope.go:117] "RemoveContainer" containerID="466d0a1c1b79a9bf590cfd803ef7a2afed5130e42ae2e0f32a8e3b35a4037bbf" Dec 09 10:13:22 crc kubenswrapper[4824]: E1209 10:13:22.567732 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-69f49fcbcf-7bnk5_openstack(ec8287ef-1c22-4524-8a7b-c20d76430002)\"" pod="openstack/heat-api-69f49fcbcf-7bnk5" podUID="ec8287ef-1c22-4524-8a7b-c20d76430002" Dec 09 10:13:23 crc kubenswrapper[4824]: I1209 10:13:23.080108 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-78488f59bc-grx2c" Dec 09 10:13:23 crc kubenswrapper[4824]: I1209 10:13:23.237801 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ddd52d5e-b6bd-403e-affe-c7d25aa4d556-config-data-custom\") pod \"ddd52d5e-b6bd-403e-affe-c7d25aa4d556\" (UID: \"ddd52d5e-b6bd-403e-affe-c7d25aa4d556\") " Dec 09 10:13:23 crc kubenswrapper[4824]: I1209 10:13:23.237862 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpfng\" (UniqueName: \"kubernetes.io/projected/ddd52d5e-b6bd-403e-affe-c7d25aa4d556-kube-api-access-hpfng\") pod \"ddd52d5e-b6bd-403e-affe-c7d25aa4d556\" (UID: \"ddd52d5e-b6bd-403e-affe-c7d25aa4d556\") " Dec 09 10:13:23 crc kubenswrapper[4824]: I1209 10:13:23.237974 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddd52d5e-b6bd-403e-affe-c7d25aa4d556-config-data\") pod \"ddd52d5e-b6bd-403e-affe-c7d25aa4d556\" (UID: \"ddd52d5e-b6bd-403e-affe-c7d25aa4d556\") " Dec 09 10:13:23 crc kubenswrapper[4824]: I1209 10:13:23.238004 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddd52d5e-b6bd-403e-affe-c7d25aa4d556-combined-ca-bundle\") pod \"ddd52d5e-b6bd-403e-affe-c7d25aa4d556\" (UID: \"ddd52d5e-b6bd-403e-affe-c7d25aa4d556\") " Dec 09 10:13:23 crc kubenswrapper[4824]: I1209 10:13:23.244258 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddd52d5e-b6bd-403e-affe-c7d25aa4d556-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ddd52d5e-b6bd-403e-affe-c7d25aa4d556" (UID: "ddd52d5e-b6bd-403e-affe-c7d25aa4d556"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:23 crc kubenswrapper[4824]: I1209 10:13:23.248023 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddd52d5e-b6bd-403e-affe-c7d25aa4d556-kube-api-access-hpfng" (OuterVolumeSpecName: "kube-api-access-hpfng") pod "ddd52d5e-b6bd-403e-affe-c7d25aa4d556" (UID: "ddd52d5e-b6bd-403e-affe-c7d25aa4d556"). InnerVolumeSpecName "kube-api-access-hpfng". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:13:23 crc kubenswrapper[4824]: I1209 10:13:23.292391 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddd52d5e-b6bd-403e-affe-c7d25aa4d556-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ddd52d5e-b6bd-403e-affe-c7d25aa4d556" (UID: "ddd52d5e-b6bd-403e-affe-c7d25aa4d556"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:23 crc kubenswrapper[4824]: I1209 10:13:23.322069 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddd52d5e-b6bd-403e-affe-c7d25aa4d556-config-data" (OuterVolumeSpecName: "config-data") pod "ddd52d5e-b6bd-403e-affe-c7d25aa4d556" (UID: "ddd52d5e-b6bd-403e-affe-c7d25aa4d556"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:23 crc kubenswrapper[4824]: I1209 10:13:23.342234 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddd52d5e-b6bd-403e-affe-c7d25aa4d556-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:23 crc kubenswrapper[4824]: I1209 10:13:23.342282 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddd52d5e-b6bd-403e-affe-c7d25aa4d556-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:23 crc kubenswrapper[4824]: I1209 10:13:23.342302 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ddd52d5e-b6bd-403e-affe-c7d25aa4d556-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:23 crc kubenswrapper[4824]: I1209 10:13:23.342318 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpfng\" (UniqueName: \"kubernetes.io/projected/ddd52d5e-b6bd-403e-affe-c7d25aa4d556-kube-api-access-hpfng\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:23 crc kubenswrapper[4824]: I1209 10:13:23.599708 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"a00596c6-582b-4ec7-9920-6dddc8197001","Type":"ContainerStarted","Data":"bd5a95612c1d9e919ee80474e8faf019afc0710bda98784f03117a6c7e9d8fb2"} Dec 09 10:13:23 crc kubenswrapper[4824]: I1209 10:13:23.603919 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-78488f59bc-grx2c" event={"ID":"ddd52d5e-b6bd-403e-affe-c7d25aa4d556","Type":"ContainerDied","Data":"69ef6324025d05aefef805521350fe9567cdd4d5cdc971040e5dd0da1b09a43f"} Dec 09 10:13:23 crc kubenswrapper[4824]: I1209 10:13:23.603986 4824 scope.go:117] "RemoveContainer" containerID="f690aba191f16dc81b5dcbfdedec49c1b3a9ddda35231e2e292c4e92691db354" Dec 09 10:13:23 crc kubenswrapper[4824]: I1209 10:13:23.604059 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-78488f59bc-grx2c" Dec 09 10:13:23 crc kubenswrapper[4824]: I1209 10:13:23.638736 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.015355022 podStartE2EDuration="38.638704862s" podCreationTimestamp="2025-12-09 10:12:45 +0000 UTC" firstStartedPulling="2025-12-09 10:12:46.962271372 +0000 UTC m=+1523.296776039" lastFinishedPulling="2025-12-09 10:13:22.585621212 +0000 UTC m=+1558.920125879" observedRunningTime="2025-12-09 10:13:23.632135742 +0000 UTC m=+1559.966640409" watchObservedRunningTime="2025-12-09 10:13:23.638704862 +0000 UTC m=+1559.973209539" Dec 09 10:13:23 crc kubenswrapper[4824]: I1209 10:13:23.720381 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-78488f59bc-grx2c"] Dec 09 10:13:23 crc kubenswrapper[4824]: I1209 10:13:23.734611 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-78488f59bc-grx2c"] Dec 09 10:13:24 crc kubenswrapper[4824]: I1209 10:13:24.078505 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddd52d5e-b6bd-403e-affe-c7d25aa4d556" path="/var/lib/kubelet/pods/ddd52d5e-b6bd-403e-affe-c7d25aa4d556/volumes" Dec 09 10:13:24 crc kubenswrapper[4824]: I1209 10:13:24.222260 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69f49fcbcf-7bnk5" Dec 09 10:13:24 crc kubenswrapper[4824]: I1209 10:13:24.313739 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzpnl\" (UniqueName: \"kubernetes.io/projected/ec8287ef-1c22-4524-8a7b-c20d76430002-kube-api-access-gzpnl\") pod \"ec8287ef-1c22-4524-8a7b-c20d76430002\" (UID: \"ec8287ef-1c22-4524-8a7b-c20d76430002\") " Dec 09 10:13:24 crc kubenswrapper[4824]: I1209 10:13:24.313890 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8287ef-1c22-4524-8a7b-c20d76430002-config-data\") pod \"ec8287ef-1c22-4524-8a7b-c20d76430002\" (UID: \"ec8287ef-1c22-4524-8a7b-c20d76430002\") " Dec 09 10:13:24 crc kubenswrapper[4824]: I1209 10:13:24.314102 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8287ef-1c22-4524-8a7b-c20d76430002-combined-ca-bundle\") pod \"ec8287ef-1c22-4524-8a7b-c20d76430002\" (UID: \"ec8287ef-1c22-4524-8a7b-c20d76430002\") " Dec 09 10:13:24 crc kubenswrapper[4824]: I1209 10:13:24.314120 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ec8287ef-1c22-4524-8a7b-c20d76430002-config-data-custom\") pod \"ec8287ef-1c22-4524-8a7b-c20d76430002\" (UID: \"ec8287ef-1c22-4524-8a7b-c20d76430002\") " Dec 09 10:13:24 crc kubenswrapper[4824]: I1209 10:13:24.321223 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec8287ef-1c22-4524-8a7b-c20d76430002-kube-api-access-gzpnl" (OuterVolumeSpecName: "kube-api-access-gzpnl") pod "ec8287ef-1c22-4524-8a7b-c20d76430002" (UID: "ec8287ef-1c22-4524-8a7b-c20d76430002"). InnerVolumeSpecName "kube-api-access-gzpnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:13:24 crc kubenswrapper[4824]: I1209 10:13:24.324934 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec8287ef-1c22-4524-8a7b-c20d76430002-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ec8287ef-1c22-4524-8a7b-c20d76430002" (UID: "ec8287ef-1c22-4524-8a7b-c20d76430002"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:24 crc kubenswrapper[4824]: I1209 10:13:24.366550 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec8287ef-1c22-4524-8a7b-c20d76430002-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec8287ef-1c22-4524-8a7b-c20d76430002" (UID: "ec8287ef-1c22-4524-8a7b-c20d76430002"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:24 crc kubenswrapper[4824]: I1209 10:13:24.393307 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec8287ef-1c22-4524-8a7b-c20d76430002-config-data" (OuterVolumeSpecName: "config-data") pod "ec8287ef-1c22-4524-8a7b-c20d76430002" (UID: "ec8287ef-1c22-4524-8a7b-c20d76430002"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:24 crc kubenswrapper[4824]: I1209 10:13:24.417306 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8287ef-1c22-4524-8a7b-c20d76430002-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:24 crc kubenswrapper[4824]: I1209 10:13:24.417356 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8287ef-1c22-4524-8a7b-c20d76430002-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:24 crc kubenswrapper[4824]: I1209 10:13:24.417373 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ec8287ef-1c22-4524-8a7b-c20d76430002-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:24 crc kubenswrapper[4824]: I1209 10:13:24.417386 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzpnl\" (UniqueName: \"kubernetes.io/projected/ec8287ef-1c22-4524-8a7b-c20d76430002-kube-api-access-gzpnl\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:24 crc kubenswrapper[4824]: I1209 10:13:24.618742 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69f49fcbcf-7bnk5" event={"ID":"ec8287ef-1c22-4524-8a7b-c20d76430002","Type":"ContainerDied","Data":"5f7c86e281ef96b58b48242f2b73732409e9e17d68ea982e6b47eee4c7c64f50"} Dec 09 10:13:24 crc kubenswrapper[4824]: I1209 10:13:24.618823 4824 scope.go:117] "RemoveContainer" containerID="466d0a1c1b79a9bf590cfd803ef7a2afed5130e42ae2e0f32a8e3b35a4037bbf" Dec 09 10:13:24 crc kubenswrapper[4824]: I1209 10:13:24.618817 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69f49fcbcf-7bnk5" Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.033924 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-69f49fcbcf-7bnk5"] Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.066930 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-69f49fcbcf-7bnk5"] Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.381616 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.515234 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9cc85a9e-8935-432d-8324-491c402402ea-sg-core-conf-yaml\") pod \"9cc85a9e-8935-432d-8324-491c402402ea\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.515339 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggbw4\" (UniqueName: \"kubernetes.io/projected/9cc85a9e-8935-432d-8324-491c402402ea-kube-api-access-ggbw4\") pod \"9cc85a9e-8935-432d-8324-491c402402ea\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.515385 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cc85a9e-8935-432d-8324-491c402402ea-combined-ca-bundle\") pod \"9cc85a9e-8935-432d-8324-491c402402ea\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.515501 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cc85a9e-8935-432d-8324-491c402402ea-log-httpd\") pod \"9cc85a9e-8935-432d-8324-491c402402ea\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.515532 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cc85a9e-8935-432d-8324-491c402402ea-scripts\") pod \"9cc85a9e-8935-432d-8324-491c402402ea\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.515637 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cc85a9e-8935-432d-8324-491c402402ea-config-data\") pod \"9cc85a9e-8935-432d-8324-491c402402ea\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.515708 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cc85a9e-8935-432d-8324-491c402402ea-run-httpd\") pod \"9cc85a9e-8935-432d-8324-491c402402ea\" (UID: \"9cc85a9e-8935-432d-8324-491c402402ea\") " Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.517273 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cc85a9e-8935-432d-8324-491c402402ea-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9cc85a9e-8935-432d-8324-491c402402ea" (UID: "9cc85a9e-8935-432d-8324-491c402402ea"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.519599 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cc85a9e-8935-432d-8324-491c402402ea-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9cc85a9e-8935-432d-8324-491c402402ea" (UID: "9cc85a9e-8935-432d-8324-491c402402ea"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.525128 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cc85a9e-8935-432d-8324-491c402402ea-scripts" (OuterVolumeSpecName: "scripts") pod "9cc85a9e-8935-432d-8324-491c402402ea" (UID: "9cc85a9e-8935-432d-8324-491c402402ea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.536060 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cc85a9e-8935-432d-8324-491c402402ea-kube-api-access-ggbw4" (OuterVolumeSpecName: "kube-api-access-ggbw4") pod "9cc85a9e-8935-432d-8324-491c402402ea" (UID: "9cc85a9e-8935-432d-8324-491c402402ea"). InnerVolumeSpecName "kube-api-access-ggbw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.566393 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cc85a9e-8935-432d-8324-491c402402ea-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9cc85a9e-8935-432d-8324-491c402402ea" (UID: "9cc85a9e-8935-432d-8324-491c402402ea"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.620929 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9cc85a9e-8935-432d-8324-491c402402ea-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.621276 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggbw4\" (UniqueName: \"kubernetes.io/projected/9cc85a9e-8935-432d-8324-491c402402ea-kube-api-access-ggbw4\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.621288 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cc85a9e-8935-432d-8324-491c402402ea-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.621296 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cc85a9e-8935-432d-8324-491c402402ea-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.621306 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cc85a9e-8935-432d-8324-491c402402ea-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.648098 4824 generic.go:334] "Generic (PLEG): container finished" podID="9cc85a9e-8935-432d-8324-491c402402ea" containerID="26dbed783b66c0024fba06c893071ac5841a36aa5722a623cd78e75bc9f4725b" exitCode=0 Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.648144 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cc85a9e-8935-432d-8324-491c402402ea","Type":"ContainerDied","Data":"26dbed783b66c0024fba06c893071ac5841a36aa5722a623cd78e75bc9f4725b"} Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.648173 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cc85a9e-8935-432d-8324-491c402402ea","Type":"ContainerDied","Data":"5564e132583313a8fd5642f07f74cc3424deeb1d6809b582acbcc982bd50ab75"} Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.648190 4824 scope.go:117] "RemoveContainer" containerID="70c869b7ae088e7888ff3f6ef92e2c78807bd104485f5bafd6f17a9276483746" Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.648313 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.700665 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cc85a9e-8935-432d-8324-491c402402ea-config-data" (OuterVolumeSpecName: "config-data") pod "9cc85a9e-8935-432d-8324-491c402402ea" (UID: "9cc85a9e-8935-432d-8324-491c402402ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.724016 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cc85a9e-8935-432d-8324-491c402402ea-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.738931 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cc85a9e-8935-432d-8324-491c402402ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9cc85a9e-8935-432d-8324-491c402402ea" (UID: "9cc85a9e-8935-432d-8324-491c402402ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.826548 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cc85a9e-8935-432d-8324-491c402402ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.849936 4824 scope.go:117] "RemoveContainer" containerID="3f141fcc82254eac5845a2bab2c48e08301ff3b94134665db9e95e6607142bb2" Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.853178 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.884367 4824 scope.go:117] "RemoveContainer" containerID="6de41f253743baba6c5372b71ee72fd1c62f8cdb896ddea6e5984165758cf820" Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.955267 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec8287ef-1c22-4524-8a7b-c20d76430002" path="/var/lib/kubelet/pods/ec8287ef-1c22-4524-8a7b-c20d76430002/volumes" Dec 09 10:13:25 crc kubenswrapper[4824]: I1209 10:13:25.963793 4824 scope.go:117] "RemoveContainer" containerID="26dbed783b66c0024fba06c893071ac5841a36aa5722a623cd78e75bc9f4725b" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.016858 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.020042 4824 scope.go:117] "RemoveContainer" containerID="70c869b7ae088e7888ff3f6ef92e2c78807bd104485f5bafd6f17a9276483746" Dec 09 10:13:26 crc kubenswrapper[4824]: E1209 10:13:26.029544 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70c869b7ae088e7888ff3f6ef92e2c78807bd104485f5bafd6f17a9276483746\": container with ID starting with 70c869b7ae088e7888ff3f6ef92e2c78807bd104485f5bafd6f17a9276483746 not found: ID does not exist" containerID="70c869b7ae088e7888ff3f6ef92e2c78807bd104485f5bafd6f17a9276483746" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.030226 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70c869b7ae088e7888ff3f6ef92e2c78807bd104485f5bafd6f17a9276483746"} err="failed to get container status \"70c869b7ae088e7888ff3f6ef92e2c78807bd104485f5bafd6f17a9276483746\": rpc error: code = NotFound desc = could not find container \"70c869b7ae088e7888ff3f6ef92e2c78807bd104485f5bafd6f17a9276483746\": container with ID starting with 70c869b7ae088e7888ff3f6ef92e2c78807bd104485f5bafd6f17a9276483746 not found: ID does not exist" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.030278 4824 scope.go:117] "RemoveContainer" containerID="3f141fcc82254eac5845a2bab2c48e08301ff3b94134665db9e95e6607142bb2" Dec 09 10:13:26 crc kubenswrapper[4824]: E1209 10:13:26.042034 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f141fcc82254eac5845a2bab2c48e08301ff3b94134665db9e95e6607142bb2\": container with ID starting with 3f141fcc82254eac5845a2bab2c48e08301ff3b94134665db9e95e6607142bb2 not found: ID does not exist" containerID="3f141fcc82254eac5845a2bab2c48e08301ff3b94134665db9e95e6607142bb2" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.042088 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f141fcc82254eac5845a2bab2c48e08301ff3b94134665db9e95e6607142bb2"} err="failed to get container status \"3f141fcc82254eac5845a2bab2c48e08301ff3b94134665db9e95e6607142bb2\": rpc error: code = NotFound desc = could not find container \"3f141fcc82254eac5845a2bab2c48e08301ff3b94134665db9e95e6607142bb2\": container with ID starting with 3f141fcc82254eac5845a2bab2c48e08301ff3b94134665db9e95e6607142bb2 not found: ID does not exist" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.042124 4824 scope.go:117] "RemoveContainer" containerID="6de41f253743baba6c5372b71ee72fd1c62f8cdb896ddea6e5984165758cf820" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.042399 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:13:26 crc kubenswrapper[4824]: E1209 10:13:26.050235 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6de41f253743baba6c5372b71ee72fd1c62f8cdb896ddea6e5984165758cf820\": container with ID starting with 6de41f253743baba6c5372b71ee72fd1c62f8cdb896ddea6e5984165758cf820 not found: ID does not exist" containerID="6de41f253743baba6c5372b71ee72fd1c62f8cdb896ddea6e5984165758cf820" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.050283 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6de41f253743baba6c5372b71ee72fd1c62f8cdb896ddea6e5984165758cf820"} err="failed to get container status \"6de41f253743baba6c5372b71ee72fd1c62f8cdb896ddea6e5984165758cf820\": rpc error: code = NotFound desc = could not find container \"6de41f253743baba6c5372b71ee72fd1c62f8cdb896ddea6e5984165758cf820\": container with ID starting with 6de41f253743baba6c5372b71ee72fd1c62f8cdb896ddea6e5984165758cf820 not found: ID does not exist" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.050316 4824 scope.go:117] "RemoveContainer" containerID="26dbed783b66c0024fba06c893071ac5841a36aa5722a623cd78e75bc9f4725b" Dec 09 10:13:26 crc kubenswrapper[4824]: E1209 10:13:26.059376 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26dbed783b66c0024fba06c893071ac5841a36aa5722a623cd78e75bc9f4725b\": container with ID starting with 26dbed783b66c0024fba06c893071ac5841a36aa5722a623cd78e75bc9f4725b not found: ID does not exist" containerID="26dbed783b66c0024fba06c893071ac5841a36aa5722a623cd78e75bc9f4725b" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.059467 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26dbed783b66c0024fba06c893071ac5841a36aa5722a623cd78e75bc9f4725b"} err="failed to get container status \"26dbed783b66c0024fba06c893071ac5841a36aa5722a623cd78e75bc9f4725b\": rpc error: code = NotFound desc = could not find container \"26dbed783b66c0024fba06c893071ac5841a36aa5722a623cd78e75bc9f4725b\": container with ID starting with 26dbed783b66c0024fba06c893071ac5841a36aa5722a623cd78e75bc9f4725b not found: ID does not exist" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.082864 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:13:26 crc kubenswrapper[4824]: E1209 10:13:26.083485 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddd52d5e-b6bd-403e-affe-c7d25aa4d556" containerName="heat-cfnapi" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.083509 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddd52d5e-b6bd-403e-affe-c7d25aa4d556" containerName="heat-cfnapi" Dec 09 10:13:26 crc kubenswrapper[4824]: E1209 10:13:26.083526 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b816dc5-829b-4c38-b2e0-b08cfbfc1617" containerName="init" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.083533 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b816dc5-829b-4c38-b2e0-b08cfbfc1617" containerName="init" Dec 09 10:13:26 crc kubenswrapper[4824]: E1209 10:13:26.083551 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec8287ef-1c22-4524-8a7b-c20d76430002" containerName="heat-api" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.083558 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec8287ef-1c22-4524-8a7b-c20d76430002" containerName="heat-api" Dec 09 10:13:26 crc kubenswrapper[4824]: E1209 10:13:26.083573 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddd52d5e-b6bd-403e-affe-c7d25aa4d556" containerName="heat-cfnapi" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.083578 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddd52d5e-b6bd-403e-affe-c7d25aa4d556" containerName="heat-cfnapi" Dec 09 10:13:26 crc kubenswrapper[4824]: E1209 10:13:26.083591 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cc85a9e-8935-432d-8324-491c402402ea" containerName="proxy-httpd" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.083597 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cc85a9e-8935-432d-8324-491c402402ea" containerName="proxy-httpd" Dec 09 10:13:26 crc kubenswrapper[4824]: E1209 10:13:26.083614 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b816dc5-829b-4c38-b2e0-b08cfbfc1617" containerName="dnsmasq-dns" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.083619 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b816dc5-829b-4c38-b2e0-b08cfbfc1617" containerName="dnsmasq-dns" Dec 09 10:13:26 crc kubenswrapper[4824]: E1209 10:13:26.083628 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cc85a9e-8935-432d-8324-491c402402ea" containerName="sg-core" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.083634 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cc85a9e-8935-432d-8324-491c402402ea" containerName="sg-core" Dec 09 10:13:26 crc kubenswrapper[4824]: E1209 10:13:26.083651 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec8287ef-1c22-4524-8a7b-c20d76430002" containerName="heat-api" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.083656 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec8287ef-1c22-4524-8a7b-c20d76430002" containerName="heat-api" Dec 09 10:13:26 crc kubenswrapper[4824]: E1209 10:13:26.083673 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cc85a9e-8935-432d-8324-491c402402ea" containerName="ceilometer-notification-agent" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.083679 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cc85a9e-8935-432d-8324-491c402402ea" containerName="ceilometer-notification-agent" Dec 09 10:13:26 crc kubenswrapper[4824]: E1209 10:13:26.083696 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cc85a9e-8935-432d-8324-491c402402ea" containerName="ceilometer-central-agent" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.083702 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cc85a9e-8935-432d-8324-491c402402ea" containerName="ceilometer-central-agent" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.083995 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cc85a9e-8935-432d-8324-491c402402ea" containerName="sg-core" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.084014 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cc85a9e-8935-432d-8324-491c402402ea" containerName="proxy-httpd" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.084024 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec8287ef-1c22-4524-8a7b-c20d76430002" containerName="heat-api" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.084032 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cc85a9e-8935-432d-8324-491c402402ea" containerName="ceilometer-notification-agent" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.084041 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddd52d5e-b6bd-403e-affe-c7d25aa4d556" containerName="heat-cfnapi" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.084052 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b816dc5-829b-4c38-b2e0-b08cfbfc1617" containerName="dnsmasq-dns" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.084062 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cc85a9e-8935-432d-8324-491c402402ea" containerName="ceilometer-central-agent" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.084075 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddd52d5e-b6bd-403e-affe-c7d25aa4d556" containerName="heat-cfnapi" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.084090 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec8287ef-1c22-4524-8a7b-c20d76430002" containerName="heat-api" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.086367 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.090177 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.091482 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.103196 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.241723 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08aaff98-7331-4684-8a51-f009911c20c8-run-httpd\") pod \"ceilometer-0\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " pod="openstack/ceilometer-0" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.242009 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08aaff98-7331-4684-8a51-f009911c20c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " pod="openstack/ceilometer-0" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.242167 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/08aaff98-7331-4684-8a51-f009911c20c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " pod="openstack/ceilometer-0" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.242200 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc4vs\" (UniqueName: \"kubernetes.io/projected/08aaff98-7331-4684-8a51-f009911c20c8-kube-api-access-wc4vs\") pod \"ceilometer-0\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " pod="openstack/ceilometer-0" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.242305 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08aaff98-7331-4684-8a51-f009911c20c8-config-data\") pod \"ceilometer-0\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " pod="openstack/ceilometer-0" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.242417 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08aaff98-7331-4684-8a51-f009911c20c8-scripts\") pod \"ceilometer-0\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " pod="openstack/ceilometer-0" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.242518 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08aaff98-7331-4684-8a51-f009911c20c8-log-httpd\") pod \"ceilometer-0\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " pod="openstack/ceilometer-0" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.344641 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08aaff98-7331-4684-8a51-f009911c20c8-run-httpd\") pod \"ceilometer-0\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " pod="openstack/ceilometer-0" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.344803 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08aaff98-7331-4684-8a51-f009911c20c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " pod="openstack/ceilometer-0" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.344864 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/08aaff98-7331-4684-8a51-f009911c20c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " pod="openstack/ceilometer-0" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.344886 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc4vs\" (UniqueName: \"kubernetes.io/projected/08aaff98-7331-4684-8a51-f009911c20c8-kube-api-access-wc4vs\") pod \"ceilometer-0\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " pod="openstack/ceilometer-0" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.344948 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08aaff98-7331-4684-8a51-f009911c20c8-config-data\") pod \"ceilometer-0\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " pod="openstack/ceilometer-0" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.345014 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08aaff98-7331-4684-8a51-f009911c20c8-scripts\") pod \"ceilometer-0\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " pod="openstack/ceilometer-0" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.345042 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08aaff98-7331-4684-8a51-f009911c20c8-log-httpd\") pod \"ceilometer-0\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " pod="openstack/ceilometer-0" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.345520 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08aaff98-7331-4684-8a51-f009911c20c8-log-httpd\") pod \"ceilometer-0\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " pod="openstack/ceilometer-0" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.345739 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08aaff98-7331-4684-8a51-f009911c20c8-run-httpd\") pod \"ceilometer-0\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " pod="openstack/ceilometer-0" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.353116 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08aaff98-7331-4684-8a51-f009911c20c8-scripts\") pod \"ceilometer-0\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " pod="openstack/ceilometer-0" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.353214 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08aaff98-7331-4684-8a51-f009911c20c8-config-data\") pod \"ceilometer-0\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " pod="openstack/ceilometer-0" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.355424 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08aaff98-7331-4684-8a51-f009911c20c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " pod="openstack/ceilometer-0" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.371921 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/08aaff98-7331-4684-8a51-f009911c20c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " pod="openstack/ceilometer-0" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.401535 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc4vs\" (UniqueName: \"kubernetes.io/projected/08aaff98-7331-4684-8a51-f009911c20c8-kube-api-access-wc4vs\") pod \"ceilometer-0\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " pod="openstack/ceilometer-0" Dec 09 10:13:26 crc kubenswrapper[4824]: I1209 10:13:26.431036 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:13:27 crc kubenswrapper[4824]: I1209 10:13:27.049338 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:13:27 crc kubenswrapper[4824]: W1209 10:13:27.059189 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08aaff98_7331_4684_8a51_f009911c20c8.slice/crio-7ffaf0e75bb2ab2cf54c0358c0573b709925d5ca0fcd7d17f50d6db0c15f0c54 WatchSource:0}: Error finding container 7ffaf0e75bb2ab2cf54c0358c0573b709925d5ca0fcd7d17f50d6db0c15f0c54: Status 404 returned error can't find the container with id 7ffaf0e75bb2ab2cf54c0358c0573b709925d5ca0fcd7d17f50d6db0c15f0c54 Dec 09 10:13:27 crc kubenswrapper[4824]: I1209 10:13:27.369404 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:13:27 crc kubenswrapper[4824]: I1209 10:13:27.706449 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08aaff98-7331-4684-8a51-f009911c20c8","Type":"ContainerStarted","Data":"7ffaf0e75bb2ab2cf54c0358c0573b709925d5ca0fcd7d17f50d6db0c15f0c54"} Dec 09 10:13:27 crc kubenswrapper[4824]: I1209 10:13:27.933464 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cc85a9e-8935-432d-8324-491c402402ea" path="/var/lib/kubelet/pods/9cc85a9e-8935-432d-8324-491c402402ea/volumes" Dec 09 10:13:28 crc kubenswrapper[4824]: I1209 10:13:28.720825 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08aaff98-7331-4684-8a51-f009911c20c8","Type":"ContainerStarted","Data":"356f5cd708b9ba96569487fc7eaad00284afc25ddc7a67736c6f22e40e529ad4"} Dec 09 10:13:28 crc kubenswrapper[4824]: I1209 10:13:28.958765 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-bzx8k"] Dec 09 10:13:28 crc kubenswrapper[4824]: I1209 10:13:28.961467 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-bzx8k" Dec 09 10:13:28 crc kubenswrapper[4824]: I1209 10:13:28.982914 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-bzx8k"] Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.041867 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-h9wzl"] Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.044065 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-h9wzl" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.058652 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-h9wzl"] Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.124095 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52e9f723-e235-4411-aeff-818c586afe9d-operator-scripts\") pod \"nova-api-db-create-bzx8k\" (UID: \"52e9f723-e235-4411-aeff-818c586afe9d\") " pod="openstack/nova-api-db-create-bzx8k" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.124249 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxnrq\" (UniqueName: \"kubernetes.io/projected/52e9f723-e235-4411-aeff-818c586afe9d-kube-api-access-vxnrq\") pod \"nova-api-db-create-bzx8k\" (UID: \"52e9f723-e235-4411-aeff-818c586afe9d\") " pod="openstack/nova-api-db-create-bzx8k" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.155457 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-840e-account-create-update-9f4xh"] Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.157643 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-840e-account-create-update-9f4xh" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.160048 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.186696 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-840e-account-create-update-9f4xh"] Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.226751 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4668f61-ef52-4072-bd2c-ed91c7298ed3-operator-scripts\") pod \"nova-cell0-db-create-h9wzl\" (UID: \"a4668f61-ef52-4072-bd2c-ed91c7298ed3\") " pod="openstack/nova-cell0-db-create-h9wzl" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.227032 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52e9f723-e235-4411-aeff-818c586afe9d-operator-scripts\") pod \"nova-api-db-create-bzx8k\" (UID: \"52e9f723-e235-4411-aeff-818c586afe9d\") " pod="openstack/nova-api-db-create-bzx8k" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.227080 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-922qp\" (UniqueName: \"kubernetes.io/projected/a4668f61-ef52-4072-bd2c-ed91c7298ed3-kube-api-access-922qp\") pod \"nova-cell0-db-create-h9wzl\" (UID: \"a4668f61-ef52-4072-bd2c-ed91c7298ed3\") " pod="openstack/nova-cell0-db-create-h9wzl" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.227157 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxnrq\" (UniqueName: \"kubernetes.io/projected/52e9f723-e235-4411-aeff-818c586afe9d-kube-api-access-vxnrq\") pod \"nova-api-db-create-bzx8k\" (UID: \"52e9f723-e235-4411-aeff-818c586afe9d\") " pod="openstack/nova-api-db-create-bzx8k" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.228268 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52e9f723-e235-4411-aeff-818c586afe9d-operator-scripts\") pod \"nova-api-db-create-bzx8k\" (UID: \"52e9f723-e235-4411-aeff-818c586afe9d\") " pod="openstack/nova-api-db-create-bzx8k" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.235837 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-2966b"] Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.238158 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2966b" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.263708 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-2966b"] Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.268380 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxnrq\" (UniqueName: \"kubernetes.io/projected/52e9f723-e235-4411-aeff-818c586afe9d-kube-api-access-vxnrq\") pod \"nova-api-db-create-bzx8k\" (UID: \"52e9f723-e235-4411-aeff-818c586afe9d\") " pod="openstack/nova-api-db-create-bzx8k" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.289383 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-bzx8k" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.335602 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-922qp\" (UniqueName: \"kubernetes.io/projected/a4668f61-ef52-4072-bd2c-ed91c7298ed3-kube-api-access-922qp\") pod \"nova-cell0-db-create-h9wzl\" (UID: \"a4668f61-ef52-4072-bd2c-ed91c7298ed3\") " pod="openstack/nova-cell0-db-create-h9wzl" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.335712 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9658659d-f0c7-47b9-95ef-14a8f4be023e-operator-scripts\") pod \"nova-api-840e-account-create-update-9f4xh\" (UID: \"9658659d-f0c7-47b9-95ef-14a8f4be023e\") " pod="openstack/nova-api-840e-account-create-update-9f4xh" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.335957 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4668f61-ef52-4072-bd2c-ed91c7298ed3-operator-scripts\") pod \"nova-cell0-db-create-h9wzl\" (UID: \"a4668f61-ef52-4072-bd2c-ed91c7298ed3\") " pod="openstack/nova-cell0-db-create-h9wzl" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.336057 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chg9g\" (UniqueName: \"kubernetes.io/projected/9658659d-f0c7-47b9-95ef-14a8f4be023e-kube-api-access-chg9g\") pod \"nova-api-840e-account-create-update-9f4xh\" (UID: \"9658659d-f0c7-47b9-95ef-14a8f4be023e\") " pod="openstack/nova-api-840e-account-create-update-9f4xh" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.337217 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4668f61-ef52-4072-bd2c-ed91c7298ed3-operator-scripts\") pod \"nova-cell0-db-create-h9wzl\" (UID: \"a4668f61-ef52-4072-bd2c-ed91c7298ed3\") " pod="openstack/nova-cell0-db-create-h9wzl" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.371427 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-922qp\" (UniqueName: \"kubernetes.io/projected/a4668f61-ef52-4072-bd2c-ed91c7298ed3-kube-api-access-922qp\") pod \"nova-cell0-db-create-h9wzl\" (UID: \"a4668f61-ef52-4072-bd2c-ed91c7298ed3\") " pod="openstack/nova-cell0-db-create-h9wzl" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.371520 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-7461-account-create-update-gvr6w"] Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.373525 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7461-account-create-update-gvr6w" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.377060 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.380830 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-h9wzl" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.423158 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-7461-account-create-update-gvr6w"] Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.438526 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6m7d\" (UniqueName: \"kubernetes.io/projected/ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee-kube-api-access-m6m7d\") pod \"nova-cell1-db-create-2966b\" (UID: \"ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee\") " pod="openstack/nova-cell1-db-create-2966b" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.439180 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee-operator-scripts\") pod \"nova-cell1-db-create-2966b\" (UID: \"ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee\") " pod="openstack/nova-cell1-db-create-2966b" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.441968 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chg9g\" (UniqueName: \"kubernetes.io/projected/9658659d-f0c7-47b9-95ef-14a8f4be023e-kube-api-access-chg9g\") pod \"nova-api-840e-account-create-update-9f4xh\" (UID: \"9658659d-f0c7-47b9-95ef-14a8f4be023e\") " pod="openstack/nova-api-840e-account-create-update-9f4xh" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.442122 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9658659d-f0c7-47b9-95ef-14a8f4be023e-operator-scripts\") pod \"nova-api-840e-account-create-update-9f4xh\" (UID: \"9658659d-f0c7-47b9-95ef-14a8f4be023e\") " pod="openstack/nova-api-840e-account-create-update-9f4xh" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.443845 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9658659d-f0c7-47b9-95ef-14a8f4be023e-operator-scripts\") pod \"nova-api-840e-account-create-update-9f4xh\" (UID: \"9658659d-f0c7-47b9-95ef-14a8f4be023e\") " pod="openstack/nova-api-840e-account-create-update-9f4xh" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.461323 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chg9g\" (UniqueName: \"kubernetes.io/projected/9658659d-f0c7-47b9-95ef-14a8f4be023e-kube-api-access-chg9g\") pod \"nova-api-840e-account-create-update-9f4xh\" (UID: \"9658659d-f0c7-47b9-95ef-14a8f4be023e\") " pod="openstack/nova-api-840e-account-create-update-9f4xh" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.477813 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-840e-account-create-update-9f4xh" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.548925 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cff7cbdd-8ed6-48e5-9fb0-35123690fac0-operator-scripts\") pod \"nova-cell0-7461-account-create-update-gvr6w\" (UID: \"cff7cbdd-8ed6-48e5-9fb0-35123690fac0\") " pod="openstack/nova-cell0-7461-account-create-update-gvr6w" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.549033 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p69w\" (UniqueName: \"kubernetes.io/projected/cff7cbdd-8ed6-48e5-9fb0-35123690fac0-kube-api-access-6p69w\") pod \"nova-cell0-7461-account-create-update-gvr6w\" (UID: \"cff7cbdd-8ed6-48e5-9fb0-35123690fac0\") " pod="openstack/nova-cell0-7461-account-create-update-gvr6w" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.549077 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6m7d\" (UniqueName: \"kubernetes.io/projected/ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee-kube-api-access-m6m7d\") pod \"nova-cell1-db-create-2966b\" (UID: \"ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee\") " pod="openstack/nova-cell1-db-create-2966b" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.549113 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee-operator-scripts\") pod \"nova-cell1-db-create-2966b\" (UID: \"ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee\") " pod="openstack/nova-cell1-db-create-2966b" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.550179 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee-operator-scripts\") pod \"nova-cell1-db-create-2966b\" (UID: \"ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee\") " pod="openstack/nova-cell1-db-create-2966b" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.569807 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-366d-account-create-update-zfvhr"] Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.571733 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-366d-account-create-update-zfvhr" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.575920 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.631538 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6m7d\" (UniqueName: \"kubernetes.io/projected/ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee-kube-api-access-m6m7d\") pod \"nova-cell1-db-create-2966b\" (UID: \"ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee\") " pod="openstack/nova-cell1-db-create-2966b" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.631614 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-366d-account-create-update-zfvhr"] Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.667994 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cff7cbdd-8ed6-48e5-9fb0-35123690fac0-operator-scripts\") pod \"nova-cell0-7461-account-create-update-gvr6w\" (UID: \"cff7cbdd-8ed6-48e5-9fb0-35123690fac0\") " pod="openstack/nova-cell0-7461-account-create-update-gvr6w" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.668114 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p69w\" (UniqueName: \"kubernetes.io/projected/cff7cbdd-8ed6-48e5-9fb0-35123690fac0-kube-api-access-6p69w\") pod \"nova-cell0-7461-account-create-update-gvr6w\" (UID: \"cff7cbdd-8ed6-48e5-9fb0-35123690fac0\") " pod="openstack/nova-cell0-7461-account-create-update-gvr6w" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.669511 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cff7cbdd-8ed6-48e5-9fb0-35123690fac0-operator-scripts\") pod \"nova-cell0-7461-account-create-update-gvr6w\" (UID: \"cff7cbdd-8ed6-48e5-9fb0-35123690fac0\") " pod="openstack/nova-cell0-7461-account-create-update-gvr6w" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.691222 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p69w\" (UniqueName: \"kubernetes.io/projected/cff7cbdd-8ed6-48e5-9fb0-35123690fac0-kube-api-access-6p69w\") pod \"nova-cell0-7461-account-create-update-gvr6w\" (UID: \"cff7cbdd-8ed6-48e5-9fb0-35123690fac0\") " pod="openstack/nova-cell0-7461-account-create-update-gvr6w" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.724355 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2966b" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.745776 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7461-account-create-update-gvr6w" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.750606 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08aaff98-7331-4684-8a51-f009911c20c8","Type":"ContainerStarted","Data":"cdba7844f942fd84a6bf5c803d335ce730f79f941f445a120a8193c83026e3bc"} Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.771380 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vdnd\" (UniqueName: \"kubernetes.io/projected/e79ef98a-aa4c-4452-b217-bbb43c73fed3-kube-api-access-7vdnd\") pod \"nova-cell1-366d-account-create-update-zfvhr\" (UID: \"e79ef98a-aa4c-4452-b217-bbb43c73fed3\") " pod="openstack/nova-cell1-366d-account-create-update-zfvhr" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.771599 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e79ef98a-aa4c-4452-b217-bbb43c73fed3-operator-scripts\") pod \"nova-cell1-366d-account-create-update-zfvhr\" (UID: \"e79ef98a-aa4c-4452-b217-bbb43c73fed3\") " pod="openstack/nova-cell1-366d-account-create-update-zfvhr" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.879424 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vdnd\" (UniqueName: \"kubernetes.io/projected/e79ef98a-aa4c-4452-b217-bbb43c73fed3-kube-api-access-7vdnd\") pod \"nova-cell1-366d-account-create-update-zfvhr\" (UID: \"e79ef98a-aa4c-4452-b217-bbb43c73fed3\") " pod="openstack/nova-cell1-366d-account-create-update-zfvhr" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.880447 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e79ef98a-aa4c-4452-b217-bbb43c73fed3-operator-scripts\") pod \"nova-cell1-366d-account-create-update-zfvhr\" (UID: \"e79ef98a-aa4c-4452-b217-bbb43c73fed3\") " pod="openstack/nova-cell1-366d-account-create-update-zfvhr" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.882293 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e79ef98a-aa4c-4452-b217-bbb43c73fed3-operator-scripts\") pod \"nova-cell1-366d-account-create-update-zfvhr\" (UID: \"e79ef98a-aa4c-4452-b217-bbb43c73fed3\") " pod="openstack/nova-cell1-366d-account-create-update-zfvhr" Dec 09 10:13:29 crc kubenswrapper[4824]: I1209 10:13:29.917900 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vdnd\" (UniqueName: \"kubernetes.io/projected/e79ef98a-aa4c-4452-b217-bbb43c73fed3-kube-api-access-7vdnd\") pod \"nova-cell1-366d-account-create-update-zfvhr\" (UID: \"e79ef98a-aa4c-4452-b217-bbb43c73fed3\") " pod="openstack/nova-cell1-366d-account-create-update-zfvhr" Dec 09 10:13:30 crc kubenswrapper[4824]: I1209 10:13:30.063224 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-366d-account-create-update-zfvhr" Dec 09 10:13:30 crc kubenswrapper[4824]: I1209 10:13:30.578208 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-h9wzl"] Dec 09 10:13:30 crc kubenswrapper[4824]: I1209 10:13:30.637851 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-840e-account-create-update-9f4xh"] Dec 09 10:13:30 crc kubenswrapper[4824]: I1209 10:13:30.646340 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-bzx8k"] Dec 09 10:13:30 crc kubenswrapper[4824]: I1209 10:13:30.886248 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08aaff98-7331-4684-8a51-f009911c20c8","Type":"ContainerStarted","Data":"bdf260dcf2266fec4424fb0cfa053c306cc4594001b912d41b973875c1c987e5"} Dec 09 10:13:30 crc kubenswrapper[4824]: I1209 10:13:30.904474 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-840e-account-create-update-9f4xh" event={"ID":"9658659d-f0c7-47b9-95ef-14a8f4be023e","Type":"ContainerStarted","Data":"7cfa3a99df77ba73f06af51346503e04117225f3e5493bc096850a8c8895791c"} Dec 09 10:13:30 crc kubenswrapper[4824]: I1209 10:13:30.933129 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-h9wzl" event={"ID":"a4668f61-ef52-4072-bd2c-ed91c7298ed3","Type":"ContainerStarted","Data":"aea35c8a6f93893297bdbd62636f8ab2bf6fa40308a3d6a6436db459fdcf77c9"} Dec 09 10:13:30 crc kubenswrapper[4824]: I1209 10:13:30.935422 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-bzx8k" event={"ID":"52e9f723-e235-4411-aeff-818c586afe9d","Type":"ContainerStarted","Data":"82ae4ee178a89f4c9efc30a27a6eb939406f3489a8f1a0eebe04c492845ec761"} Dec 09 10:13:31 crc kubenswrapper[4824]: I1209 10:13:31.171736 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-7461-account-create-update-gvr6w"] Dec 09 10:13:31 crc kubenswrapper[4824]: I1209 10:13:31.262348 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-2966b"] Dec 09 10:13:31 crc kubenswrapper[4824]: I1209 10:13:31.389506 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-366d-account-create-update-zfvhr"] Dec 09 10:13:31 crc kubenswrapper[4824]: I1209 10:13:31.967323 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-366d-account-create-update-zfvhr" event={"ID":"e79ef98a-aa4c-4452-b217-bbb43c73fed3","Type":"ContainerStarted","Data":"4eca3ba123a696c5c1bc56ff873888e3dfc0e19a54bca8d5315eae93e7f40e4a"} Dec 09 10:13:31 crc kubenswrapper[4824]: I1209 10:13:31.981698 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-840e-account-create-update-9f4xh" event={"ID":"9658659d-f0c7-47b9-95ef-14a8f4be023e","Type":"ContainerStarted","Data":"77650c36dee46cfe0e7fdbcbd89b23ea80a5e7ae6ea5ba7c73f25de9c50b2f2c"} Dec 09 10:13:31 crc kubenswrapper[4824]: I1209 10:13:31.995008 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-h9wzl" event={"ID":"a4668f61-ef52-4072-bd2c-ed91c7298ed3","Type":"ContainerStarted","Data":"525064878d70a2f6e2ffadc3478f6b6c99c7a46e197c0d269b691fc42ec86995"} Dec 09 10:13:32 crc kubenswrapper[4824]: I1209 10:13:32.012544 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-2966b" event={"ID":"ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee","Type":"ContainerStarted","Data":"336fd1f138c5c4d8b275c5dcd43c629a890af24e4c6f66067dd236913df49747"} Dec 09 10:13:32 crc kubenswrapper[4824]: I1209 10:13:32.037025 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-bzx8k" event={"ID":"52e9f723-e235-4411-aeff-818c586afe9d","Type":"ContainerStarted","Data":"a6e8c0b9da4a4795c1e6d82834746a67795b93a047b9b70e2cc2c184147b7013"} Dec 09 10:13:32 crc kubenswrapper[4824]: I1209 10:13:32.037960 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-840e-account-create-update-9f4xh" podStartSLOduration=3.037938425 podStartE2EDuration="3.037938425s" podCreationTimestamp="2025-12-09 10:13:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:13:32.007408713 +0000 UTC m=+1568.341913390" watchObservedRunningTime="2025-12-09 10:13:32.037938425 +0000 UTC m=+1568.372443092" Dec 09 10:13:32 crc kubenswrapper[4824]: I1209 10:13:32.038858 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-h9wzl" podStartSLOduration=3.038852043 podStartE2EDuration="3.038852043s" podCreationTimestamp="2025-12-09 10:13:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:13:32.024989062 +0000 UTC m=+1568.359493729" watchObservedRunningTime="2025-12-09 10:13:32.038852043 +0000 UTC m=+1568.373356710" Dec 09 10:13:32 crc kubenswrapper[4824]: I1209 10:13:32.048057 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7461-account-create-update-gvr6w" event={"ID":"cff7cbdd-8ed6-48e5-9fb0-35123690fac0","Type":"ContainerStarted","Data":"141239e059fdb05a5e2d0171af0d3165bff8e5a6556f9784e61da5c2a8ee55e0"} Dec 09 10:13:32 crc kubenswrapper[4824]: I1209 10:13:32.061672 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-7d54cc949d-rzfx4" Dec 09 10:13:32 crc kubenswrapper[4824]: I1209 10:13:32.067416 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-bzx8k" podStartSLOduration=4.067390002 podStartE2EDuration="4.067390002s" podCreationTimestamp="2025-12-09 10:13:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:13:32.055466372 +0000 UTC m=+1568.389971049" watchObservedRunningTime="2025-12-09 10:13:32.067390002 +0000 UTC m=+1568.401894659" Dec 09 10:13:32 crc kubenswrapper[4824]: I1209 10:13:32.146004 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-6fffc89c95-wt6ms"] Dec 09 10:13:32 crc kubenswrapper[4824]: I1209 10:13:32.146402 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-6fffc89c95-wt6ms" podUID="04bcf8d0-9439-4739-a503-e504651a80dd" containerName="heat-engine" containerID="cri-o://5b1cb02d98415e13ebd728dc83bdef09545646e7c9bbe6f2b829ceebe2c6ab0f" gracePeriod=60 Dec 09 10:13:33 crc kubenswrapper[4824]: I1209 10:13:33.063164 4824 generic.go:334] "Generic (PLEG): container finished" podID="e79ef98a-aa4c-4452-b217-bbb43c73fed3" containerID="1f146c01309ed0552d65c2763666643b8d29faf00a88737fd4ba628b79c8cd9e" exitCode=0 Dec 09 10:13:33 crc kubenswrapper[4824]: I1209 10:13:33.063210 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-366d-account-create-update-zfvhr" event={"ID":"e79ef98a-aa4c-4452-b217-bbb43c73fed3","Type":"ContainerDied","Data":"1f146c01309ed0552d65c2763666643b8d29faf00a88737fd4ba628b79c8cd9e"} Dec 09 10:13:33 crc kubenswrapper[4824]: I1209 10:13:33.066191 4824 generic.go:334] "Generic (PLEG): container finished" podID="9658659d-f0c7-47b9-95ef-14a8f4be023e" containerID="77650c36dee46cfe0e7fdbcbd89b23ea80a5e7ae6ea5ba7c73f25de9c50b2f2c" exitCode=0 Dec 09 10:13:33 crc kubenswrapper[4824]: I1209 10:13:33.066278 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-840e-account-create-update-9f4xh" event={"ID":"9658659d-f0c7-47b9-95ef-14a8f4be023e","Type":"ContainerDied","Data":"77650c36dee46cfe0e7fdbcbd89b23ea80a5e7ae6ea5ba7c73f25de9c50b2f2c"} Dec 09 10:13:33 crc kubenswrapper[4824]: I1209 10:13:33.068669 4824 generic.go:334] "Generic (PLEG): container finished" podID="a4668f61-ef52-4072-bd2c-ed91c7298ed3" containerID="525064878d70a2f6e2ffadc3478f6b6c99c7a46e197c0d269b691fc42ec86995" exitCode=0 Dec 09 10:13:33 crc kubenswrapper[4824]: I1209 10:13:33.068879 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-h9wzl" event={"ID":"a4668f61-ef52-4072-bd2c-ed91c7298ed3","Type":"ContainerDied","Data":"525064878d70a2f6e2ffadc3478f6b6c99c7a46e197c0d269b691fc42ec86995"} Dec 09 10:13:33 crc kubenswrapper[4824]: I1209 10:13:33.071801 4824 generic.go:334] "Generic (PLEG): container finished" podID="ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee" containerID="bc239bc9d807eae04960e5f9ed3cb500ed60adef14d484a2d64e327fee158164" exitCode=0 Dec 09 10:13:33 crc kubenswrapper[4824]: I1209 10:13:33.071829 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-2966b" event={"ID":"ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee","Type":"ContainerDied","Data":"bc239bc9d807eae04960e5f9ed3cb500ed60adef14d484a2d64e327fee158164"} Dec 09 10:13:33 crc kubenswrapper[4824]: I1209 10:13:33.073731 4824 generic.go:334] "Generic (PLEG): container finished" podID="52e9f723-e235-4411-aeff-818c586afe9d" containerID="a6e8c0b9da4a4795c1e6d82834746a67795b93a047b9b70e2cc2c184147b7013" exitCode=0 Dec 09 10:13:33 crc kubenswrapper[4824]: I1209 10:13:33.073848 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-bzx8k" event={"ID":"52e9f723-e235-4411-aeff-818c586afe9d","Type":"ContainerDied","Data":"a6e8c0b9da4a4795c1e6d82834746a67795b93a047b9b70e2cc2c184147b7013"} Dec 09 10:13:33 crc kubenswrapper[4824]: I1209 10:13:33.075617 4824 generic.go:334] "Generic (PLEG): container finished" podID="cff7cbdd-8ed6-48e5-9fb0-35123690fac0" containerID="b252953e0c9dd129c85c067821d00db94fe90ad5cad332b4f2c3fb877c7d20bb" exitCode=0 Dec 09 10:13:33 crc kubenswrapper[4824]: I1209 10:13:33.075898 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7461-account-create-update-gvr6w" event={"ID":"cff7cbdd-8ed6-48e5-9fb0-35123690fac0","Type":"ContainerDied","Data":"b252953e0c9dd129c85c067821d00db94fe90ad5cad332b4f2c3fb877c7d20bb"} Dec 09 10:13:33 crc kubenswrapper[4824]: I1209 10:13:33.078600 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08aaff98-7331-4684-8a51-f009911c20c8","Type":"ContainerStarted","Data":"5c951d74cbe52d50d259dc7c9f784a070563d7903262cb67cd1ffd8bfa4fa207"} Dec 09 10:13:33 crc kubenswrapper[4824]: I1209 10:13:33.078951 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="08aaff98-7331-4684-8a51-f009911c20c8" containerName="ceilometer-central-agent" containerID="cri-o://356f5cd708b9ba96569487fc7eaad00284afc25ddc7a67736c6f22e40e529ad4" gracePeriod=30 Dec 09 10:13:33 crc kubenswrapper[4824]: I1209 10:13:33.079171 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 10:13:33 crc kubenswrapper[4824]: I1209 10:13:33.079311 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="08aaff98-7331-4684-8a51-f009911c20c8" containerName="proxy-httpd" containerID="cri-o://5c951d74cbe52d50d259dc7c9f784a070563d7903262cb67cd1ffd8bfa4fa207" gracePeriod=30 Dec 09 10:13:33 crc kubenswrapper[4824]: I1209 10:13:33.079371 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="08aaff98-7331-4684-8a51-f009911c20c8" containerName="sg-core" containerID="cri-o://bdf260dcf2266fec4424fb0cfa053c306cc4594001b912d41b973875c1c987e5" gracePeriod=30 Dec 09 10:13:33 crc kubenswrapper[4824]: I1209 10:13:33.079470 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="08aaff98-7331-4684-8a51-f009911c20c8" containerName="ceilometer-notification-agent" containerID="cri-o://cdba7844f942fd84a6bf5c803d335ce730f79f941f445a120a8193c83026e3bc" gracePeriod=30 Dec 09 10:13:33 crc kubenswrapper[4824]: I1209 10:13:33.164124 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.6813548369999998 podStartE2EDuration="7.164092909s" podCreationTimestamp="2025-12-09 10:13:26 +0000 UTC" firstStartedPulling="2025-12-09 10:13:27.062970834 +0000 UTC m=+1563.397475501" lastFinishedPulling="2025-12-09 10:13:31.545708906 +0000 UTC m=+1567.880213573" observedRunningTime="2025-12-09 10:13:33.158719378 +0000 UTC m=+1569.493224055" watchObservedRunningTime="2025-12-09 10:13:33.164092909 +0000 UTC m=+1569.498597596" Dec 09 10:13:33 crc kubenswrapper[4824]: E1209 10:13:33.531796 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08aaff98_7331_4684_8a51_f009911c20c8.slice/crio-bdf260dcf2266fec4424fb0cfa053c306cc4594001b912d41b973875c1c987e5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08aaff98_7331_4684_8a51_f009911c20c8.slice/crio-conmon-bdf260dcf2266fec4424fb0cfa053c306cc4594001b912d41b973875c1c987e5.scope\": RecentStats: unable to find data in memory cache]" Dec 09 10:13:33 crc kubenswrapper[4824]: I1209 10:13:33.906441 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 10:13:33 crc kubenswrapper[4824]: I1209 10:13:33.907078 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d974a3d8-22a8-4301-bb5f-28011189b4dc" containerName="glance-log" containerID="cri-o://ada896b2969c2cfab19365a8fe216e3c2acd2174978d17ef9a24d6c1b93b028d" gracePeriod=30 Dec 09 10:13:33 crc kubenswrapper[4824]: I1209 10:13:33.907225 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d974a3d8-22a8-4301-bb5f-28011189b4dc" containerName="glance-httpd" containerID="cri-o://0763c351bab138a499fba2d8c867605e7f3926bfe2ca508f44cd60ce8bb1b6d9" gracePeriod=30 Dec 09 10:13:34 crc kubenswrapper[4824]: I1209 10:13:34.094483 4824 generic.go:334] "Generic (PLEG): container finished" podID="d974a3d8-22a8-4301-bb5f-28011189b4dc" containerID="ada896b2969c2cfab19365a8fe216e3c2acd2174978d17ef9a24d6c1b93b028d" exitCode=143 Dec 09 10:13:34 crc kubenswrapper[4824]: I1209 10:13:34.094575 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d974a3d8-22a8-4301-bb5f-28011189b4dc","Type":"ContainerDied","Data":"ada896b2969c2cfab19365a8fe216e3c2acd2174978d17ef9a24d6c1b93b028d"} Dec 09 10:13:34 crc kubenswrapper[4824]: I1209 10:13:34.097974 4824 generic.go:334] "Generic (PLEG): container finished" podID="08aaff98-7331-4684-8a51-f009911c20c8" containerID="bdf260dcf2266fec4424fb0cfa053c306cc4594001b912d41b973875c1c987e5" exitCode=2 Dec 09 10:13:34 crc kubenswrapper[4824]: I1209 10:13:34.098039 4824 generic.go:334] "Generic (PLEG): container finished" podID="08aaff98-7331-4684-8a51-f009911c20c8" containerID="cdba7844f942fd84a6bf5c803d335ce730f79f941f445a120a8193c83026e3bc" exitCode=0 Dec 09 10:13:34 crc kubenswrapper[4824]: I1209 10:13:34.098057 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08aaff98-7331-4684-8a51-f009911c20c8","Type":"ContainerDied","Data":"bdf260dcf2266fec4424fb0cfa053c306cc4594001b912d41b973875c1c987e5"} Dec 09 10:13:34 crc kubenswrapper[4824]: I1209 10:13:34.098121 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08aaff98-7331-4684-8a51-f009911c20c8","Type":"ContainerDied","Data":"cdba7844f942fd84a6bf5c803d335ce730f79f941f445a120a8193c83026e3bc"} Dec 09 10:13:34 crc kubenswrapper[4824]: I1209 10:13:34.669356 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-366d-account-create-update-zfvhr" Dec 09 10:13:34 crc kubenswrapper[4824]: I1209 10:13:34.826156 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e79ef98a-aa4c-4452-b217-bbb43c73fed3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e79ef98a-aa4c-4452-b217-bbb43c73fed3" (UID: "e79ef98a-aa4c-4452-b217-bbb43c73fed3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:13:34 crc kubenswrapper[4824]: I1209 10:13:34.826683 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e79ef98a-aa4c-4452-b217-bbb43c73fed3-operator-scripts\") pod \"e79ef98a-aa4c-4452-b217-bbb43c73fed3\" (UID: \"e79ef98a-aa4c-4452-b217-bbb43c73fed3\") " Dec 09 10:13:34 crc kubenswrapper[4824]: I1209 10:13:34.827058 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vdnd\" (UniqueName: \"kubernetes.io/projected/e79ef98a-aa4c-4452-b217-bbb43c73fed3-kube-api-access-7vdnd\") pod \"e79ef98a-aa4c-4452-b217-bbb43c73fed3\" (UID: \"e79ef98a-aa4c-4452-b217-bbb43c73fed3\") " Dec 09 10:13:34 crc kubenswrapper[4824]: I1209 10:13:34.831011 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e79ef98a-aa4c-4452-b217-bbb43c73fed3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:34 crc kubenswrapper[4824]: I1209 10:13:34.859142 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e79ef98a-aa4c-4452-b217-bbb43c73fed3-kube-api-access-7vdnd" (OuterVolumeSpecName: "kube-api-access-7vdnd") pod "e79ef98a-aa4c-4452-b217-bbb43c73fed3" (UID: "e79ef98a-aa4c-4452-b217-bbb43c73fed3"). InnerVolumeSpecName "kube-api-access-7vdnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:13:34 crc kubenswrapper[4824]: I1209 10:13:34.945600 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vdnd\" (UniqueName: \"kubernetes.io/projected/e79ef98a-aa4c-4452-b217-bbb43c73fed3-kube-api-access-7vdnd\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.107809 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-bzx8k" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.129105 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-840e-account-create-update-9f4xh" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.129330 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-366d-account-create-update-zfvhr" event={"ID":"e79ef98a-aa4c-4452-b217-bbb43c73fed3","Type":"ContainerDied","Data":"4eca3ba123a696c5c1bc56ff873888e3dfc0e19a54bca8d5315eae93e7f40e4a"} Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.129367 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4eca3ba123a696c5c1bc56ff873888e3dfc0e19a54bca8d5315eae93e7f40e4a" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.129421 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-366d-account-create-update-zfvhr" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.144449 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-840e-account-create-update-9f4xh" event={"ID":"9658659d-f0c7-47b9-95ef-14a8f4be023e","Type":"ContainerDied","Data":"7cfa3a99df77ba73f06af51346503e04117225f3e5493bc096850a8c8895791c"} Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.144497 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7cfa3a99df77ba73f06af51346503e04117225f3e5493bc096850a8c8895791c" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.144589 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-840e-account-create-update-9f4xh" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.162346 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-h9wzl" event={"ID":"a4668f61-ef52-4072-bd2c-ed91c7298ed3","Type":"ContainerDied","Data":"aea35c8a6f93893297bdbd62636f8ab2bf6fa40308a3d6a6436db459fdcf77c9"} Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.162399 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aea35c8a6f93893297bdbd62636f8ab2bf6fa40308a3d6a6436db459fdcf77c9" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.166432 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-2966b" event={"ID":"ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee","Type":"ContainerDied","Data":"336fd1f138c5c4d8b275c5dcd43c629a890af24e4c6f66067dd236913df49747"} Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.166470 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="336fd1f138c5c4d8b275c5dcd43c629a890af24e4c6f66067dd236913df49747" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.168137 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-bzx8k" event={"ID":"52e9f723-e235-4411-aeff-818c586afe9d","Type":"ContainerDied","Data":"82ae4ee178a89f4c9efc30a27a6eb939406f3489a8f1a0eebe04c492845ec761"} Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.168168 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82ae4ee178a89f4c9efc30a27a6eb939406f3489a8f1a0eebe04c492845ec761" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.168245 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-bzx8k" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.177196 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7461-account-create-update-gvr6w" event={"ID":"cff7cbdd-8ed6-48e5-9fb0-35123690fac0","Type":"ContainerDied","Data":"141239e059fdb05a5e2d0171af0d3165bff8e5a6556f9784e61da5c2a8ee55e0"} Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.177239 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="141239e059fdb05a5e2d0171af0d3165bff8e5a6556f9784e61da5c2a8ee55e0" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.201155 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-h9wzl" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.210215 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2966b" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.232627 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7461-account-create-update-gvr6w" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.260643 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chg9g\" (UniqueName: \"kubernetes.io/projected/9658659d-f0c7-47b9-95ef-14a8f4be023e-kube-api-access-chg9g\") pod \"9658659d-f0c7-47b9-95ef-14a8f4be023e\" (UID: \"9658659d-f0c7-47b9-95ef-14a8f4be023e\") " Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.260716 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4668f61-ef52-4072-bd2c-ed91c7298ed3-operator-scripts\") pod \"a4668f61-ef52-4072-bd2c-ed91c7298ed3\" (UID: \"a4668f61-ef52-4072-bd2c-ed91c7298ed3\") " Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.263866 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-922qp\" (UniqueName: \"kubernetes.io/projected/a4668f61-ef52-4072-bd2c-ed91c7298ed3-kube-api-access-922qp\") pod \"a4668f61-ef52-4072-bd2c-ed91c7298ed3\" (UID: \"a4668f61-ef52-4072-bd2c-ed91c7298ed3\") " Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.263917 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxnrq\" (UniqueName: \"kubernetes.io/projected/52e9f723-e235-4411-aeff-818c586afe9d-kube-api-access-vxnrq\") pod \"52e9f723-e235-4411-aeff-818c586afe9d\" (UID: \"52e9f723-e235-4411-aeff-818c586afe9d\") " Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.264033 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52e9f723-e235-4411-aeff-818c586afe9d-operator-scripts\") pod \"52e9f723-e235-4411-aeff-818c586afe9d\" (UID: \"52e9f723-e235-4411-aeff-818c586afe9d\") " Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.264080 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9658659d-f0c7-47b9-95ef-14a8f4be023e-operator-scripts\") pod \"9658659d-f0c7-47b9-95ef-14a8f4be023e\" (UID: \"9658659d-f0c7-47b9-95ef-14a8f4be023e\") " Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.268602 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52e9f723-e235-4411-aeff-818c586afe9d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "52e9f723-e235-4411-aeff-818c586afe9d" (UID: "52e9f723-e235-4411-aeff-818c586afe9d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.273094 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4668f61-ef52-4072-bd2c-ed91c7298ed3-kube-api-access-922qp" (OuterVolumeSpecName: "kube-api-access-922qp") pod "a4668f61-ef52-4072-bd2c-ed91c7298ed3" (UID: "a4668f61-ef52-4072-bd2c-ed91c7298ed3"). InnerVolumeSpecName "kube-api-access-922qp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.273429 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9658659d-f0c7-47b9-95ef-14a8f4be023e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9658659d-f0c7-47b9-95ef-14a8f4be023e" (UID: "9658659d-f0c7-47b9-95ef-14a8f4be023e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.273870 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4668f61-ef52-4072-bd2c-ed91c7298ed3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a4668f61-ef52-4072-bd2c-ed91c7298ed3" (UID: "a4668f61-ef52-4072-bd2c-ed91c7298ed3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.278283 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52e9f723-e235-4411-aeff-818c586afe9d-kube-api-access-vxnrq" (OuterVolumeSpecName: "kube-api-access-vxnrq") pod "52e9f723-e235-4411-aeff-818c586afe9d" (UID: "52e9f723-e235-4411-aeff-818c586afe9d"). InnerVolumeSpecName "kube-api-access-vxnrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.278857 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9658659d-f0c7-47b9-95ef-14a8f4be023e-kube-api-access-chg9g" (OuterVolumeSpecName: "kube-api-access-chg9g") pod "9658659d-f0c7-47b9-95ef-14a8f4be023e" (UID: "9658659d-f0c7-47b9-95ef-14a8f4be023e"). InnerVolumeSpecName "kube-api-access-chg9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.368698 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee-operator-scripts\") pod \"ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee\" (UID: \"ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee\") " Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.368948 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cff7cbdd-8ed6-48e5-9fb0-35123690fac0-operator-scripts\") pod \"cff7cbdd-8ed6-48e5-9fb0-35123690fac0\" (UID: \"cff7cbdd-8ed6-48e5-9fb0-35123690fac0\") " Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.369110 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6m7d\" (UniqueName: \"kubernetes.io/projected/ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee-kube-api-access-m6m7d\") pod \"ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee\" (UID: \"ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee\") " Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.369224 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6p69w\" (UniqueName: \"kubernetes.io/projected/cff7cbdd-8ed6-48e5-9fb0-35123690fac0-kube-api-access-6p69w\") pod \"cff7cbdd-8ed6-48e5-9fb0-35123690fac0\" (UID: \"cff7cbdd-8ed6-48e5-9fb0-35123690fac0\") " Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.369955 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9658659d-f0c7-47b9-95ef-14a8f4be023e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.369975 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chg9g\" (UniqueName: \"kubernetes.io/projected/9658659d-f0c7-47b9-95ef-14a8f4be023e-kube-api-access-chg9g\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.370049 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4668f61-ef52-4072-bd2c-ed91c7298ed3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.370062 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-922qp\" (UniqueName: \"kubernetes.io/projected/a4668f61-ef52-4072-bd2c-ed91c7298ed3-kube-api-access-922qp\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.370074 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxnrq\" (UniqueName: \"kubernetes.io/projected/52e9f723-e235-4411-aeff-818c586afe9d-kube-api-access-vxnrq\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.370087 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52e9f723-e235-4411-aeff-818c586afe9d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.372559 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cff7cbdd-8ed6-48e5-9fb0-35123690fac0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cff7cbdd-8ed6-48e5-9fb0-35123690fac0" (UID: "cff7cbdd-8ed6-48e5-9fb0-35123690fac0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.373861 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee" (UID: "ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.375467 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cff7cbdd-8ed6-48e5-9fb0-35123690fac0-kube-api-access-6p69w" (OuterVolumeSpecName: "kube-api-access-6p69w") pod "cff7cbdd-8ed6-48e5-9fb0-35123690fac0" (UID: "cff7cbdd-8ed6-48e5-9fb0-35123690fac0"). InnerVolumeSpecName "kube-api-access-6p69w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.377601 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee-kube-api-access-m6m7d" (OuterVolumeSpecName: "kube-api-access-m6m7d") pod "ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee" (UID: "ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee"). InnerVolumeSpecName "kube-api-access-m6m7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.472970 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cff7cbdd-8ed6-48e5-9fb0-35123690fac0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.473219 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6m7d\" (UniqueName: \"kubernetes.io/projected/ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee-kube-api-access-m6m7d\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.473333 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6p69w\" (UniqueName: \"kubernetes.io/projected/cff7cbdd-8ed6-48e5-9fb0-35123690fac0-kube-api-access-6p69w\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:35 crc kubenswrapper[4824]: I1209 10:13:35.473436 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:36 crc kubenswrapper[4824]: I1209 10:13:36.198372 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-h9wzl" Dec 09 10:13:36 crc kubenswrapper[4824]: I1209 10:13:36.199200 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7461-account-create-update-gvr6w" Dec 09 10:13:36 crc kubenswrapper[4824]: I1209 10:13:36.199851 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2966b" Dec 09 10:13:36 crc kubenswrapper[4824]: I1209 10:13:36.737974 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 10:13:36 crc kubenswrapper[4824]: I1209 10:13:36.738555 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="988197de-9531-4cae-92d0-ec7f836a8c73" containerName="glance-log" containerID="cri-o://75fa1185360d66e483556e52bfc7505304668e9f6690cfd56a3faffca30c8311" gracePeriod=30 Dec 09 10:13:36 crc kubenswrapper[4824]: I1209 10:13:36.738604 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="988197de-9531-4cae-92d0-ec7f836a8c73" containerName="glance-httpd" containerID="cri-o://a2aedab1b4a2c8aabfcc925c11aee95fa4283be5ad4e6682922565866d35a7c2" gracePeriod=30 Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.246431 4824 generic.go:334] "Generic (PLEG): container finished" podID="988197de-9531-4cae-92d0-ec7f836a8c73" containerID="75fa1185360d66e483556e52bfc7505304668e9f6690cfd56a3faffca30c8311" exitCode=143 Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.246923 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"988197de-9531-4cae-92d0-ec7f836a8c73","Type":"ContainerDied","Data":"75fa1185360d66e483556e52bfc7505304668e9f6690cfd56a3faffca30c8311"} Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.273167 4824 generic.go:334] "Generic (PLEG): container finished" podID="d974a3d8-22a8-4301-bb5f-28011189b4dc" containerID="0763c351bab138a499fba2d8c867605e7f3926bfe2ca508f44cd60ce8bb1b6d9" exitCode=0 Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.273225 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d974a3d8-22a8-4301-bb5f-28011189b4dc","Type":"ContainerDied","Data":"0763c351bab138a499fba2d8c867605e7f3926bfe2ca508f44cd60ce8bb1b6d9"} Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.784072 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.848897 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\") pod \"d974a3d8-22a8-4301-bb5f-28011189b4dc\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.848972 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8lf2\" (UniqueName: \"kubernetes.io/projected/d974a3d8-22a8-4301-bb5f-28011189b4dc-kube-api-access-g8lf2\") pod \"d974a3d8-22a8-4301-bb5f-28011189b4dc\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.849057 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d974a3d8-22a8-4301-bb5f-28011189b4dc-httpd-run\") pod \"d974a3d8-22a8-4301-bb5f-28011189b4dc\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.849234 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d974a3d8-22a8-4301-bb5f-28011189b4dc-config-data\") pod \"d974a3d8-22a8-4301-bb5f-28011189b4dc\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.849358 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d974a3d8-22a8-4301-bb5f-28011189b4dc-combined-ca-bundle\") pod \"d974a3d8-22a8-4301-bb5f-28011189b4dc\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.849447 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d974a3d8-22a8-4301-bb5f-28011189b4dc-logs\") pod \"d974a3d8-22a8-4301-bb5f-28011189b4dc\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.849492 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d974a3d8-22a8-4301-bb5f-28011189b4dc-scripts\") pod \"d974a3d8-22a8-4301-bb5f-28011189b4dc\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.852421 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d974a3d8-22a8-4301-bb5f-28011189b4dc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d974a3d8-22a8-4301-bb5f-28011189b4dc" (UID: "d974a3d8-22a8-4301-bb5f-28011189b4dc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.852717 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d974a3d8-22a8-4301-bb5f-28011189b4dc-logs" (OuterVolumeSpecName: "logs") pod "d974a3d8-22a8-4301-bb5f-28011189b4dc" (UID: "d974a3d8-22a8-4301-bb5f-28011189b4dc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.861862 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d974a3d8-22a8-4301-bb5f-28011189b4dc-scripts" (OuterVolumeSpecName: "scripts") pod "d974a3d8-22a8-4301-bb5f-28011189b4dc" (UID: "d974a3d8-22a8-4301-bb5f-28011189b4dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.887988 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d974a3d8-22a8-4301-bb5f-28011189b4dc-kube-api-access-g8lf2" (OuterVolumeSpecName: "kube-api-access-g8lf2") pod "d974a3d8-22a8-4301-bb5f-28011189b4dc" (UID: "d974a3d8-22a8-4301-bb5f-28011189b4dc"). InnerVolumeSpecName "kube-api-access-g8lf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.901373 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d974a3d8-22a8-4301-bb5f-28011189b4dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d974a3d8-22a8-4301-bb5f-28011189b4dc" (UID: "d974a3d8-22a8-4301-bb5f-28011189b4dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.910031 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69" (OuterVolumeSpecName: "glance") pod "d974a3d8-22a8-4301-bb5f-28011189b4dc" (UID: "d974a3d8-22a8-4301-bb5f-28011189b4dc"). InnerVolumeSpecName "pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.953487 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d974a3d8-22a8-4301-bb5f-28011189b4dc-internal-tls-certs\") pod \"d974a3d8-22a8-4301-bb5f-28011189b4dc\" (UID: \"d974a3d8-22a8-4301-bb5f-28011189b4dc\") " Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.955699 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d974a3d8-22a8-4301-bb5f-28011189b4dc-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.955739 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d974a3d8-22a8-4301-bb5f-28011189b4dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.955754 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d974a3d8-22a8-4301-bb5f-28011189b4dc-logs\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.955764 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d974a3d8-22a8-4301-bb5f-28011189b4dc-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.955828 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\") on node \"crc\" " Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.955844 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8lf2\" (UniqueName: \"kubernetes.io/projected/d974a3d8-22a8-4301-bb5f-28011189b4dc-kube-api-access-g8lf2\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:37 crc kubenswrapper[4824]: I1209 10:13:37.972000 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d974a3d8-22a8-4301-bb5f-28011189b4dc-config-data" (OuterVolumeSpecName: "config-data") pod "d974a3d8-22a8-4301-bb5f-28011189b4dc" (UID: "d974a3d8-22a8-4301-bb5f-28011189b4dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.035290 4824 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.035720 4824 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69") on node "crc" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.060507 4824 reconciler_common.go:293] "Volume detached for volume \"pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.061518 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d974a3d8-22a8-4301-bb5f-28011189b4dc-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.067391 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d974a3d8-22a8-4301-bb5f-28011189b4dc-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d974a3d8-22a8-4301-bb5f-28011189b4dc" (UID: "d974a3d8-22a8-4301-bb5f-28011189b4dc"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.164811 4824 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d974a3d8-22a8-4301-bb5f-28011189b4dc-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.315327 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d974a3d8-22a8-4301-bb5f-28011189b4dc","Type":"ContainerDied","Data":"f2bdb66743ef7568d69f8bfe9763eeb74d67d2d41cadafb0196db3222cc2275d"} Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.315395 4824 scope.go:117] "RemoveContainer" containerID="0763c351bab138a499fba2d8c867605e7f3926bfe2ca508f44cd60ce8bb1b6d9" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.315583 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.358284 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.371248 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.397439 4824 scope.go:117] "RemoveContainer" containerID="ada896b2969c2cfab19365a8fe216e3c2acd2174978d17ef9a24d6c1b93b028d" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.405234 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 10:13:38 crc kubenswrapper[4824]: E1209 10:13:38.405907 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee" containerName="mariadb-database-create" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.405939 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee" containerName="mariadb-database-create" Dec 09 10:13:38 crc kubenswrapper[4824]: E1209 10:13:38.405956 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cff7cbdd-8ed6-48e5-9fb0-35123690fac0" containerName="mariadb-account-create-update" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.405964 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cff7cbdd-8ed6-48e5-9fb0-35123690fac0" containerName="mariadb-account-create-update" Dec 09 10:13:38 crc kubenswrapper[4824]: E1209 10:13:38.405978 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d974a3d8-22a8-4301-bb5f-28011189b4dc" containerName="glance-log" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.405985 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d974a3d8-22a8-4301-bb5f-28011189b4dc" containerName="glance-log" Dec 09 10:13:38 crc kubenswrapper[4824]: E1209 10:13:38.406021 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52e9f723-e235-4411-aeff-818c586afe9d" containerName="mariadb-database-create" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.406029 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="52e9f723-e235-4411-aeff-818c586afe9d" containerName="mariadb-database-create" Dec 09 10:13:38 crc kubenswrapper[4824]: E1209 10:13:38.406048 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9658659d-f0c7-47b9-95ef-14a8f4be023e" containerName="mariadb-account-create-update" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.406054 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9658659d-f0c7-47b9-95ef-14a8f4be023e" containerName="mariadb-account-create-update" Dec 09 10:13:38 crc kubenswrapper[4824]: E1209 10:13:38.406093 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e79ef98a-aa4c-4452-b217-bbb43c73fed3" containerName="mariadb-account-create-update" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.406101 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e79ef98a-aa4c-4452-b217-bbb43c73fed3" containerName="mariadb-account-create-update" Dec 09 10:13:38 crc kubenswrapper[4824]: E1209 10:13:38.406112 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d974a3d8-22a8-4301-bb5f-28011189b4dc" containerName="glance-httpd" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.406117 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d974a3d8-22a8-4301-bb5f-28011189b4dc" containerName="glance-httpd" Dec 09 10:13:38 crc kubenswrapper[4824]: E1209 10:13:38.406129 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4668f61-ef52-4072-bd2c-ed91c7298ed3" containerName="mariadb-database-create" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.406135 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4668f61-ef52-4072-bd2c-ed91c7298ed3" containerName="mariadb-database-create" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.406349 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="52e9f723-e235-4411-aeff-818c586afe9d" containerName="mariadb-database-create" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.406363 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee" containerName="mariadb-database-create" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.406376 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d974a3d8-22a8-4301-bb5f-28011189b4dc" containerName="glance-httpd" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.406387 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="cff7cbdd-8ed6-48e5-9fb0-35123690fac0" containerName="mariadb-account-create-update" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.406397 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e79ef98a-aa4c-4452-b217-bbb43c73fed3" containerName="mariadb-account-create-update" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.406404 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9658659d-f0c7-47b9-95ef-14a8f4be023e" containerName="mariadb-account-create-update" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.406418 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4668f61-ef52-4072-bd2c-ed91c7298ed3" containerName="mariadb-database-create" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.406439 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d974a3d8-22a8-4301-bb5f-28011189b4dc" containerName="glance-log" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.408083 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.412282 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.413328 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.437934 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.473669 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a1be792-ec64-4666-ada9-1ca72712213f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.473836 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zxwg\" (UniqueName: \"kubernetes.io/projected/5a1be792-ec64-4666-ada9-1ca72712213f-kube-api-access-6zxwg\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.473888 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a1be792-ec64-4666-ada9-1ca72712213f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.473920 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a1be792-ec64-4666-ada9-1ca72712213f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.473956 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.474004 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a1be792-ec64-4666-ada9-1ca72712213f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.474036 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a1be792-ec64-4666-ada9-1ca72712213f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.474116 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a1be792-ec64-4666-ada9-1ca72712213f-logs\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.576610 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a1be792-ec64-4666-ada9-1ca72712213f-logs\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.577136 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a1be792-ec64-4666-ada9-1ca72712213f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.577186 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a1be792-ec64-4666-ada9-1ca72712213f-logs\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.577384 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zxwg\" (UniqueName: \"kubernetes.io/projected/5a1be792-ec64-4666-ada9-1ca72712213f-kube-api-access-6zxwg\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.577481 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a1be792-ec64-4666-ada9-1ca72712213f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.577539 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a1be792-ec64-4666-ada9-1ca72712213f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.577597 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.577685 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a1be792-ec64-4666-ada9-1ca72712213f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.577721 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a1be792-ec64-4666-ada9-1ca72712213f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.578746 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a1be792-ec64-4666-ada9-1ca72712213f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.582926 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a1be792-ec64-4666-ada9-1ca72712213f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.583981 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a1be792-ec64-4666-ada9-1ca72712213f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.584614 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a1be792-ec64-4666-ada9-1ca72712213f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.588303 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a1be792-ec64-4666-ada9-1ca72712213f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.608254 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zxwg\" (UniqueName: \"kubernetes.io/projected/5a1be792-ec64-4666-ada9-1ca72712213f-kube-api-access-6zxwg\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.634907 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.634969 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/50da71ebd66a0437f4c3596841b2bf86ae7bfd6dc1ba487324809d1cf1c5f125/globalmount\"" pod="openstack/glance-default-internal-api-0" Dec 09 10:13:38 crc kubenswrapper[4824]: I1209 10:13:38.849314 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca715c0c-d6dd-4f1d-9237-660ce55efd69\") pod \"glance-default-internal-api-0\" (UID: \"5a1be792-ec64-4666-ada9-1ca72712213f\") " pod="openstack/glance-default-internal-api-0" Dec 09 10:13:39 crc kubenswrapper[4824]: I1209 10:13:39.050635 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 09 10:13:39 crc kubenswrapper[4824]: I1209 10:13:39.689760 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8q4b9"] Dec 09 10:13:39 crc kubenswrapper[4824]: I1209 10:13:39.692609 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8q4b9" Dec 09 10:13:39 crc kubenswrapper[4824]: I1209 10:13:39.695348 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 09 10:13:39 crc kubenswrapper[4824]: I1209 10:13:39.695540 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 09 10:13:39 crc kubenswrapper[4824]: I1209 10:13:39.707897 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-8t2cn" Dec 09 10:13:39 crc kubenswrapper[4824]: I1209 10:13:39.712196 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8q4b9"] Dec 09 10:13:39 crc kubenswrapper[4824]: I1209 10:13:39.784228 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 10:13:39 crc kubenswrapper[4824]: I1209 10:13:39.815934 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/221367da-a9dd-488d-a2c5-69bc0f57e589-scripts\") pod \"nova-cell0-conductor-db-sync-8q4b9\" (UID: \"221367da-a9dd-488d-a2c5-69bc0f57e589\") " pod="openstack/nova-cell0-conductor-db-sync-8q4b9" Dec 09 10:13:39 crc kubenswrapper[4824]: I1209 10:13:39.816062 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngt5h\" (UniqueName: \"kubernetes.io/projected/221367da-a9dd-488d-a2c5-69bc0f57e589-kube-api-access-ngt5h\") pod \"nova-cell0-conductor-db-sync-8q4b9\" (UID: \"221367da-a9dd-488d-a2c5-69bc0f57e589\") " pod="openstack/nova-cell0-conductor-db-sync-8q4b9" Dec 09 10:13:39 crc kubenswrapper[4824]: I1209 10:13:39.816281 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/221367da-a9dd-488d-a2c5-69bc0f57e589-config-data\") pod \"nova-cell0-conductor-db-sync-8q4b9\" (UID: \"221367da-a9dd-488d-a2c5-69bc0f57e589\") " pod="openstack/nova-cell0-conductor-db-sync-8q4b9" Dec 09 10:13:39 crc kubenswrapper[4824]: I1209 10:13:39.816504 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/221367da-a9dd-488d-a2c5-69bc0f57e589-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-8q4b9\" (UID: \"221367da-a9dd-488d-a2c5-69bc0f57e589\") " pod="openstack/nova-cell0-conductor-db-sync-8q4b9" Dec 09 10:13:39 crc kubenswrapper[4824]: I1209 10:13:39.920313 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/221367da-a9dd-488d-a2c5-69bc0f57e589-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-8q4b9\" (UID: \"221367da-a9dd-488d-a2c5-69bc0f57e589\") " pod="openstack/nova-cell0-conductor-db-sync-8q4b9" Dec 09 10:13:39 crc kubenswrapper[4824]: I1209 10:13:39.920388 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/221367da-a9dd-488d-a2c5-69bc0f57e589-scripts\") pod \"nova-cell0-conductor-db-sync-8q4b9\" (UID: \"221367da-a9dd-488d-a2c5-69bc0f57e589\") " pod="openstack/nova-cell0-conductor-db-sync-8q4b9" Dec 09 10:13:39 crc kubenswrapper[4824]: I1209 10:13:39.920473 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngt5h\" (UniqueName: \"kubernetes.io/projected/221367da-a9dd-488d-a2c5-69bc0f57e589-kube-api-access-ngt5h\") pod \"nova-cell0-conductor-db-sync-8q4b9\" (UID: \"221367da-a9dd-488d-a2c5-69bc0f57e589\") " pod="openstack/nova-cell0-conductor-db-sync-8q4b9" Dec 09 10:13:39 crc kubenswrapper[4824]: I1209 10:13:39.920586 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/221367da-a9dd-488d-a2c5-69bc0f57e589-config-data\") pod \"nova-cell0-conductor-db-sync-8q4b9\" (UID: \"221367da-a9dd-488d-a2c5-69bc0f57e589\") " pod="openstack/nova-cell0-conductor-db-sync-8q4b9" Dec 09 10:13:39 crc kubenswrapper[4824]: I1209 10:13:39.927935 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/221367da-a9dd-488d-a2c5-69bc0f57e589-scripts\") pod \"nova-cell0-conductor-db-sync-8q4b9\" (UID: \"221367da-a9dd-488d-a2c5-69bc0f57e589\") " pod="openstack/nova-cell0-conductor-db-sync-8q4b9" Dec 09 10:13:39 crc kubenswrapper[4824]: I1209 10:13:39.931312 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/221367da-a9dd-488d-a2c5-69bc0f57e589-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-8q4b9\" (UID: \"221367da-a9dd-488d-a2c5-69bc0f57e589\") " pod="openstack/nova-cell0-conductor-db-sync-8q4b9" Dec 09 10:13:39 crc kubenswrapper[4824]: I1209 10:13:39.932085 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/221367da-a9dd-488d-a2c5-69bc0f57e589-config-data\") pod \"nova-cell0-conductor-db-sync-8q4b9\" (UID: \"221367da-a9dd-488d-a2c5-69bc0f57e589\") " pod="openstack/nova-cell0-conductor-db-sync-8q4b9" Dec 09 10:13:39 crc kubenswrapper[4824]: I1209 10:13:39.950622 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d974a3d8-22a8-4301-bb5f-28011189b4dc" path="/var/lib/kubelet/pods/d974a3d8-22a8-4301-bb5f-28011189b4dc/volumes" Dec 09 10:13:39 crc kubenswrapper[4824]: I1209 10:13:39.955298 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngt5h\" (UniqueName: \"kubernetes.io/projected/221367da-a9dd-488d-a2c5-69bc0f57e589-kube-api-access-ngt5h\") pod \"nova-cell0-conductor-db-sync-8q4b9\" (UID: \"221367da-a9dd-488d-a2c5-69bc0f57e589\") " pod="openstack/nova-cell0-conductor-db-sync-8q4b9" Dec 09 10:13:40 crc kubenswrapper[4824]: I1209 10:13:40.031767 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8q4b9" Dec 09 10:13:40 crc kubenswrapper[4824]: I1209 10:13:40.396260 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5a1be792-ec64-4666-ada9-1ca72712213f","Type":"ContainerStarted","Data":"372a85311d0fcbeca054e1bf616ce32364763abccde5ae1367737fd6d031e8cb"} Dec 09 10:13:40 crc kubenswrapper[4824]: I1209 10:13:40.794863 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8q4b9"] Dec 09 10:13:41 crc kubenswrapper[4824]: E1209 10:13:41.013252 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5b1cb02d98415e13ebd728dc83bdef09545646e7c9bbe6f2b829ceebe2c6ab0f" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 09 10:13:41 crc kubenswrapper[4824]: E1209 10:13:41.028453 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5b1cb02d98415e13ebd728dc83bdef09545646e7c9bbe6f2b829ceebe2c6ab0f" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 09 10:13:41 crc kubenswrapper[4824]: E1209 10:13:41.032767 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5b1cb02d98415e13ebd728dc83bdef09545646e7c9bbe6f2b829ceebe2c6ab0f" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 09 10:13:41 crc kubenswrapper[4824]: E1209 10:13:41.033015 4824 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-6fffc89c95-wt6ms" podUID="04bcf8d0-9439-4739-a503-e504651a80dd" containerName="heat-engine" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.213829 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.273319 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/988197de-9531-4cae-92d0-ec7f836a8c73-public-tls-certs\") pod \"988197de-9531-4cae-92d0-ec7f836a8c73\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.275520 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/988197de-9531-4cae-92d0-ec7f836a8c73-scripts\") pod \"988197de-9531-4cae-92d0-ec7f836a8c73\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.275614 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/988197de-9531-4cae-92d0-ec7f836a8c73-config-data\") pod \"988197de-9531-4cae-92d0-ec7f836a8c73\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.275692 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvlqq\" (UniqueName: \"kubernetes.io/projected/988197de-9531-4cae-92d0-ec7f836a8c73-kube-api-access-bvlqq\") pod \"988197de-9531-4cae-92d0-ec7f836a8c73\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.275865 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/988197de-9531-4cae-92d0-ec7f836a8c73-combined-ca-bundle\") pod \"988197de-9531-4cae-92d0-ec7f836a8c73\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.275970 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/988197de-9531-4cae-92d0-ec7f836a8c73-httpd-run\") pod \"988197de-9531-4cae-92d0-ec7f836a8c73\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.276025 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/988197de-9531-4cae-92d0-ec7f836a8c73-logs\") pod \"988197de-9531-4cae-92d0-ec7f836a8c73\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.283002 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\") pod \"988197de-9531-4cae-92d0-ec7f836a8c73\" (UID: \"988197de-9531-4cae-92d0-ec7f836a8c73\") " Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.286127 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/988197de-9531-4cae-92d0-ec7f836a8c73-logs" (OuterVolumeSpecName: "logs") pod "988197de-9531-4cae-92d0-ec7f836a8c73" (UID: "988197de-9531-4cae-92d0-ec7f836a8c73"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.291170 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/988197de-9531-4cae-92d0-ec7f836a8c73-scripts" (OuterVolumeSpecName: "scripts") pod "988197de-9531-4cae-92d0-ec7f836a8c73" (UID: "988197de-9531-4cae-92d0-ec7f836a8c73"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.292540 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/988197de-9531-4cae-92d0-ec7f836a8c73-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "988197de-9531-4cae-92d0-ec7f836a8c73" (UID: "988197de-9531-4cae-92d0-ec7f836a8c73"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.298357 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/988197de-9531-4cae-92d0-ec7f836a8c73-kube-api-access-bvlqq" (OuterVolumeSpecName: "kube-api-access-bvlqq") pod "988197de-9531-4cae-92d0-ec7f836a8c73" (UID: "988197de-9531-4cae-92d0-ec7f836a8c73"). InnerVolumeSpecName "kube-api-access-bvlqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.332466 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd" (OuterVolumeSpecName: "glance") pod "988197de-9531-4cae-92d0-ec7f836a8c73" (UID: "988197de-9531-4cae-92d0-ec7f836a8c73"). InnerVolumeSpecName "pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.379249 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/988197de-9531-4cae-92d0-ec7f836a8c73-config-data" (OuterVolumeSpecName: "config-data") pod "988197de-9531-4cae-92d0-ec7f836a8c73" (UID: "988197de-9531-4cae-92d0-ec7f836a8c73"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.386863 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/988197de-9531-4cae-92d0-ec7f836a8c73-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.386897 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/988197de-9531-4cae-92d0-ec7f836a8c73-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.386910 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvlqq\" (UniqueName: \"kubernetes.io/projected/988197de-9531-4cae-92d0-ec7f836a8c73-kube-api-access-bvlqq\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.386922 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/988197de-9531-4cae-92d0-ec7f836a8c73-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.386933 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/988197de-9531-4cae-92d0-ec7f836a8c73-logs\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.386963 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\") on node \"crc\" " Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.396419 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/988197de-9531-4cae-92d0-ec7f836a8c73-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "988197de-9531-4cae-92d0-ec7f836a8c73" (UID: "988197de-9531-4cae-92d0-ec7f836a8c73"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.415997 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/988197de-9531-4cae-92d0-ec7f836a8c73-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "988197de-9531-4cae-92d0-ec7f836a8c73" (UID: "988197de-9531-4cae-92d0-ec7f836a8c73"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.418937 4824 generic.go:334] "Generic (PLEG): container finished" podID="988197de-9531-4cae-92d0-ec7f836a8c73" containerID="a2aedab1b4a2c8aabfcc925c11aee95fa4283be5ad4e6682922565866d35a7c2" exitCode=0 Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.419188 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"988197de-9531-4cae-92d0-ec7f836a8c73","Type":"ContainerDied","Data":"a2aedab1b4a2c8aabfcc925c11aee95fa4283be5ad4e6682922565866d35a7c2"} Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.419533 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"988197de-9531-4cae-92d0-ec7f836a8c73","Type":"ContainerDied","Data":"6f3e45ac8a89ab64150f1214fbb1c24d6bbd24b99cf2811f5f28ec7c5d7d341b"} Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.419734 4824 scope.go:117] "RemoveContainer" containerID="a2aedab1b4a2c8aabfcc925c11aee95fa4283be5ad4e6682922565866d35a7c2" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.422936 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.461912 4824 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.462115 4824 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd") on node "crc" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.489796 4824 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/988197de-9531-4cae-92d0-ec7f836a8c73-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.489931 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/988197de-9531-4cae-92d0-ec7f836a8c73-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.490035 4824 reconciler_common.go:293] "Volume detached for volume \"pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.492651 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8q4b9" event={"ID":"221367da-a9dd-488d-a2c5-69bc0f57e589","Type":"ContainerStarted","Data":"784eac1cecb8dce7d30876abcd3edcdfb80bd51c6d28ba85a0238a0690e39a59"} Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.496079 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5a1be792-ec64-4666-ada9-1ca72712213f","Type":"ContainerStarted","Data":"5e7ae58d700d6f5c046320c0b4af5b14408b46f1fa77f345cccb143a61b3eb0f"} Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.503991 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.532447 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.553139 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 10:13:41 crc kubenswrapper[4824]: E1209 10:13:41.553879 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="988197de-9531-4cae-92d0-ec7f836a8c73" containerName="glance-log" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.553905 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="988197de-9531-4cae-92d0-ec7f836a8c73" containerName="glance-log" Dec 09 10:13:41 crc kubenswrapper[4824]: E1209 10:13:41.553980 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="988197de-9531-4cae-92d0-ec7f836a8c73" containerName="glance-httpd" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.553989 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="988197de-9531-4cae-92d0-ec7f836a8c73" containerName="glance-httpd" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.554280 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="988197de-9531-4cae-92d0-ec7f836a8c73" containerName="glance-httpd" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.554307 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="988197de-9531-4cae-92d0-ec7f836a8c73" containerName="glance-log" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.556087 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.568094 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.574701 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.593464 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.608210 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22dc07da-136d-4812-b6c4-b7fa94903b5e-scripts\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") " pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.608391 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") " pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.608472 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22dc07da-136d-4812-b6c4-b7fa94903b5e-logs\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") " pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.608493 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/22dc07da-136d-4812-b6c4-b7fa94903b5e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") " pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.608541 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/22dc07da-136d-4812-b6c4-b7fa94903b5e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") " pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.608575 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22dc07da-136d-4812-b6c4-b7fa94903b5e-config-data\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") " pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.608593 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdvnq\" (UniqueName: \"kubernetes.io/projected/22dc07da-136d-4812-b6c4-b7fa94903b5e-kube-api-access-xdvnq\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") " pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.608702 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22dc07da-136d-4812-b6c4-b7fa94903b5e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") " pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.633098 4824 scope.go:117] "RemoveContainer" containerID="75fa1185360d66e483556e52bfc7505304668e9f6690cfd56a3faffca30c8311" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.706239 4824 scope.go:117] "RemoveContainer" containerID="a2aedab1b4a2c8aabfcc925c11aee95fa4283be5ad4e6682922565866d35a7c2" Dec 09 10:13:41 crc kubenswrapper[4824]: E1209 10:13:41.712064 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2aedab1b4a2c8aabfcc925c11aee95fa4283be5ad4e6682922565866d35a7c2\": container with ID starting with a2aedab1b4a2c8aabfcc925c11aee95fa4283be5ad4e6682922565866d35a7c2 not found: ID does not exist" containerID="a2aedab1b4a2c8aabfcc925c11aee95fa4283be5ad4e6682922565866d35a7c2" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.712144 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2aedab1b4a2c8aabfcc925c11aee95fa4283be5ad4e6682922565866d35a7c2"} err="failed to get container status \"a2aedab1b4a2c8aabfcc925c11aee95fa4283be5ad4e6682922565866d35a7c2\": rpc error: code = NotFound desc = could not find container \"a2aedab1b4a2c8aabfcc925c11aee95fa4283be5ad4e6682922565866d35a7c2\": container with ID starting with a2aedab1b4a2c8aabfcc925c11aee95fa4283be5ad4e6682922565866d35a7c2 not found: ID does not exist" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.712183 4824 scope.go:117] "RemoveContainer" containerID="75fa1185360d66e483556e52bfc7505304668e9f6690cfd56a3faffca30c8311" Dec 09 10:13:41 crc kubenswrapper[4824]: E1209 10:13:41.713723 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75fa1185360d66e483556e52bfc7505304668e9f6690cfd56a3faffca30c8311\": container with ID starting with 75fa1185360d66e483556e52bfc7505304668e9f6690cfd56a3faffca30c8311 not found: ID does not exist" containerID="75fa1185360d66e483556e52bfc7505304668e9f6690cfd56a3faffca30c8311" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.713804 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75fa1185360d66e483556e52bfc7505304668e9f6690cfd56a3faffca30c8311"} err="failed to get container status \"75fa1185360d66e483556e52bfc7505304668e9f6690cfd56a3faffca30c8311\": rpc error: code = NotFound desc = could not find container \"75fa1185360d66e483556e52bfc7505304668e9f6690cfd56a3faffca30c8311\": container with ID starting with 75fa1185360d66e483556e52bfc7505304668e9f6690cfd56a3faffca30c8311 not found: ID does not exist" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.717249 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") " pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.717424 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22dc07da-136d-4812-b6c4-b7fa94903b5e-logs\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") " pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.717470 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/22dc07da-136d-4812-b6c4-b7fa94903b5e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") " pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.717591 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/22dc07da-136d-4812-b6c4-b7fa94903b5e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") " pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.717672 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22dc07da-136d-4812-b6c4-b7fa94903b5e-config-data\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") " pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.717698 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdvnq\" (UniqueName: \"kubernetes.io/projected/22dc07da-136d-4812-b6c4-b7fa94903b5e-kube-api-access-xdvnq\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") " pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.718133 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22dc07da-136d-4812-b6c4-b7fa94903b5e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") " pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.718487 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22dc07da-136d-4812-b6c4-b7fa94903b5e-scripts\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") " pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.720111 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22dc07da-136d-4812-b6c4-b7fa94903b5e-logs\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") " pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.721216 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/22dc07da-136d-4812-b6c4-b7fa94903b5e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") " pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.729510 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.729899 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/afe41c6e88a0ab9650f976db82a109be2feb044bddc024f73824c502bea72164/globalmount\"" pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.731073 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/22dc07da-136d-4812-b6c4-b7fa94903b5e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") " pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.732671 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22dc07da-136d-4812-b6c4-b7fa94903b5e-scripts\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") " pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.733114 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22dc07da-136d-4812-b6c4-b7fa94903b5e-config-data\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") " pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.734194 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22dc07da-136d-4812-b6c4-b7fa94903b5e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") " pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.765635 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdvnq\" (UniqueName: \"kubernetes.io/projected/22dc07da-136d-4812-b6c4-b7fa94903b5e-kube-api-access-xdvnq\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") " pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.856277 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ffddc1d2-7447-49f6-b1eb-14ae5de17ffd\") pod \"glance-default-external-api-0\" (UID: \"22dc07da-136d-4812-b6c4-b7fa94903b5e\") " pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.897433 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 09 10:13:41 crc kubenswrapper[4824]: I1209 10:13:41.938183 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="988197de-9531-4cae-92d0-ec7f836a8c73" path="/var/lib/kubelet/pods/988197de-9531-4cae-92d0-ec7f836a8c73/volumes" Dec 09 10:13:42 crc kubenswrapper[4824]: I1209 10:13:42.520895 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5a1be792-ec64-4666-ada9-1ca72712213f","Type":"ContainerStarted","Data":"494658ef3cc1e08243546910223615468e12d28a79d19bde7afcf691a3d285bf"} Dec 09 10:13:42 crc kubenswrapper[4824]: I1209 10:13:42.556395 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.556371349 podStartE2EDuration="4.556371349s" podCreationTimestamp="2025-12-09 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:13:42.547026353 +0000 UTC m=+1578.881531030" watchObservedRunningTime="2025-12-09 10:13:42.556371349 +0000 UTC m=+1578.890876016" Dec 09 10:13:43 crc kubenswrapper[4824]: I1209 10:13:42.711645 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 10:13:43 crc kubenswrapper[4824]: W1209 10:13:42.722926 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22dc07da_136d_4812_b6c4_b7fa94903b5e.slice/crio-7f5f4724ed9e816053b83413689f68b56044815c078fec15afbaf96c672690a6 WatchSource:0}: Error finding container 7f5f4724ed9e816053b83413689f68b56044815c078fec15afbaf96c672690a6: Status 404 returned error can't find the container with id 7f5f4724ed9e816053b83413689f68b56044815c078fec15afbaf96c672690a6 Dec 09 10:13:43 crc kubenswrapper[4824]: I1209 10:13:43.541288 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"22dc07da-136d-4812-b6c4-b7fa94903b5e","Type":"ContainerStarted","Data":"3cf2c247b3ec076175856d068a7d3f6dc022c2d0337bb54c0ff2207e870fd42e"} Dec 09 10:13:43 crc kubenswrapper[4824]: I1209 10:13:43.542643 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"22dc07da-136d-4812-b6c4-b7fa94903b5e","Type":"ContainerStarted","Data":"7f5f4724ed9e816053b83413689f68b56044815c078fec15afbaf96c672690a6"} Dec 09 10:13:44 crc kubenswrapper[4824]: I1209 10:13:44.582932 4824 generic.go:334] "Generic (PLEG): container finished" podID="04bcf8d0-9439-4739-a503-e504651a80dd" containerID="5b1cb02d98415e13ebd728dc83bdef09545646e7c9bbe6f2b829ceebe2c6ab0f" exitCode=0 Dec 09 10:13:44 crc kubenswrapper[4824]: I1209 10:13:44.583214 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6fffc89c95-wt6ms" event={"ID":"04bcf8d0-9439-4739-a503-e504651a80dd","Type":"ContainerDied","Data":"5b1cb02d98415e13ebd728dc83bdef09545646e7c9bbe6f2b829ceebe2c6ab0f"} Dec 09 10:13:44 crc kubenswrapper[4824]: I1209 10:13:44.583594 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6fffc89c95-wt6ms" event={"ID":"04bcf8d0-9439-4739-a503-e504651a80dd","Type":"ContainerDied","Data":"ad73bbffb835406c8227acbaa03f03e4ddfc98ccf531b5e7a4e00a36acc7794a"} Dec 09 10:13:44 crc kubenswrapper[4824]: I1209 10:13:44.583618 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad73bbffb835406c8227acbaa03f03e4ddfc98ccf531b5e7a4e00a36acc7794a" Dec 09 10:13:44 crc kubenswrapper[4824]: I1209 10:13:44.590878 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"22dc07da-136d-4812-b6c4-b7fa94903b5e","Type":"ContainerStarted","Data":"5713f4092f1651a2f886af3f1646997e7b8cdf6dbc51fbee2d823ae2fb2c70da"} Dec 09 10:13:44 crc kubenswrapper[4824]: I1209 10:13:44.632341 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.632318266 podStartE2EDuration="3.632318266s" podCreationTimestamp="2025-12-09 10:13:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:13:44.63211493 +0000 UTC m=+1580.966619627" watchObservedRunningTime="2025-12-09 10:13:44.632318266 +0000 UTC m=+1580.966822933" Dec 09 10:13:44 crc kubenswrapper[4824]: I1209 10:13:44.688476 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6fffc89c95-wt6ms" Dec 09 10:13:44 crc kubenswrapper[4824]: I1209 10:13:44.841353 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04bcf8d0-9439-4739-a503-e504651a80dd-combined-ca-bundle\") pod \"04bcf8d0-9439-4739-a503-e504651a80dd\" (UID: \"04bcf8d0-9439-4739-a503-e504651a80dd\") " Dec 09 10:13:44 crc kubenswrapper[4824]: I1209 10:13:44.841412 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04bcf8d0-9439-4739-a503-e504651a80dd-config-data-custom\") pod \"04bcf8d0-9439-4739-a503-e504651a80dd\" (UID: \"04bcf8d0-9439-4739-a503-e504651a80dd\") " Dec 09 10:13:44 crc kubenswrapper[4824]: I1209 10:13:44.841506 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27qpm\" (UniqueName: \"kubernetes.io/projected/04bcf8d0-9439-4739-a503-e504651a80dd-kube-api-access-27qpm\") pod \"04bcf8d0-9439-4739-a503-e504651a80dd\" (UID: \"04bcf8d0-9439-4739-a503-e504651a80dd\") " Dec 09 10:13:44 crc kubenswrapper[4824]: I1209 10:13:44.841561 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04bcf8d0-9439-4739-a503-e504651a80dd-config-data\") pod \"04bcf8d0-9439-4739-a503-e504651a80dd\" (UID: \"04bcf8d0-9439-4739-a503-e504651a80dd\") " Dec 09 10:13:44 crc kubenswrapper[4824]: I1209 10:13:44.848929 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04bcf8d0-9439-4739-a503-e504651a80dd-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "04bcf8d0-9439-4739-a503-e504651a80dd" (UID: "04bcf8d0-9439-4739-a503-e504651a80dd"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:44 crc kubenswrapper[4824]: I1209 10:13:44.852127 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04bcf8d0-9439-4739-a503-e504651a80dd-kube-api-access-27qpm" (OuterVolumeSpecName: "kube-api-access-27qpm") pod "04bcf8d0-9439-4739-a503-e504651a80dd" (UID: "04bcf8d0-9439-4739-a503-e504651a80dd"). InnerVolumeSpecName "kube-api-access-27qpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:13:44 crc kubenswrapper[4824]: I1209 10:13:44.908001 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04bcf8d0-9439-4739-a503-e504651a80dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04bcf8d0-9439-4739-a503-e504651a80dd" (UID: "04bcf8d0-9439-4739-a503-e504651a80dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:44 crc kubenswrapper[4824]: I1209 10:13:44.944220 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04bcf8d0-9439-4739-a503-e504651a80dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:44 crc kubenswrapper[4824]: I1209 10:13:44.944445 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04bcf8d0-9439-4739-a503-e504651a80dd-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:44 crc kubenswrapper[4824]: I1209 10:13:44.944502 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27qpm\" (UniqueName: \"kubernetes.io/projected/04bcf8d0-9439-4739-a503-e504651a80dd-kube-api-access-27qpm\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:44 crc kubenswrapper[4824]: I1209 10:13:44.954595 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04bcf8d0-9439-4739-a503-e504651a80dd-config-data" (OuterVolumeSpecName: "config-data") pod "04bcf8d0-9439-4739-a503-e504651a80dd" (UID: "04bcf8d0-9439-4739-a503-e504651a80dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:13:45 crc kubenswrapper[4824]: I1209 10:13:45.048027 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04bcf8d0-9439-4739-a503-e504651a80dd-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:13:45 crc kubenswrapper[4824]: I1209 10:13:45.613901 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6fffc89c95-wt6ms" Dec 09 10:13:45 crc kubenswrapper[4824]: I1209 10:13:45.654259 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-6fffc89c95-wt6ms"] Dec 09 10:13:45 crc kubenswrapper[4824]: I1209 10:13:45.666529 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-6fffc89c95-wt6ms"] Dec 09 10:13:45 crc kubenswrapper[4824]: I1209 10:13:45.931527 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04bcf8d0-9439-4739-a503-e504651a80dd" path="/var/lib/kubelet/pods/04bcf8d0-9439-4739-a503-e504651a80dd/volumes" Dec 09 10:13:49 crc kubenswrapper[4824]: I1209 10:13:49.051868 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 09 10:13:49 crc kubenswrapper[4824]: I1209 10:13:49.052537 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 09 10:13:49 crc kubenswrapper[4824]: I1209 10:13:49.110333 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 09 10:13:49 crc kubenswrapper[4824]: I1209 10:13:49.120674 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 09 10:13:49 crc kubenswrapper[4824]: I1209 10:13:49.670885 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 09 10:13:49 crc kubenswrapper[4824]: I1209 10:13:49.671306 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 09 10:13:50 crc kubenswrapper[4824]: I1209 10:13:50.696489 4824 generic.go:334] "Generic (PLEG): container finished" podID="08aaff98-7331-4684-8a51-f009911c20c8" containerID="356f5cd708b9ba96569487fc7eaad00284afc25ddc7a67736c6f22e40e529ad4" exitCode=0 Dec 09 10:13:50 crc kubenswrapper[4824]: I1209 10:13:50.696542 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08aaff98-7331-4684-8a51-f009911c20c8","Type":"ContainerDied","Data":"356f5cd708b9ba96569487fc7eaad00284afc25ddc7a67736c6f22e40e529ad4"} Dec 09 10:13:51 crc kubenswrapper[4824]: I1209 10:13:51.898434 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 09 10:13:51 crc kubenswrapper[4824]: I1209 10:13:51.898745 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 09 10:13:51 crc kubenswrapper[4824]: I1209 10:13:51.950040 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 09 10:13:51 crc kubenswrapper[4824]: I1209 10:13:51.953341 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 09 10:13:52 crc kubenswrapper[4824]: I1209 10:13:52.707373 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 09 10:13:52 crc kubenswrapper[4824]: I1209 10:13:52.707470 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 10:13:52 crc kubenswrapper[4824]: I1209 10:13:52.715799 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 09 10:13:52 crc kubenswrapper[4824]: I1209 10:13:52.719593 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 09 10:13:52 crc kubenswrapper[4824]: I1209 10:13:52.719634 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 09 10:13:55 crc kubenswrapper[4824]: I1209 10:13:55.727035 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 09 10:13:55 crc kubenswrapper[4824]: I1209 10:13:55.727926 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 10:13:55 crc kubenswrapper[4824]: I1209 10:13:55.759761 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 09 10:13:55 crc kubenswrapper[4824]: I1209 10:13:55.775289 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8q4b9" event={"ID":"221367da-a9dd-488d-a2c5-69bc0f57e589","Type":"ContainerStarted","Data":"70b7050442fdf6eb10381f437e45fa00761a5e37ce90432c54bcefc8cd9db7b2"} Dec 09 10:13:55 crc kubenswrapper[4824]: I1209 10:13:55.807902 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-8q4b9" podStartSLOduration=2.719611697 podStartE2EDuration="16.807877438s" podCreationTimestamp="2025-12-09 10:13:39 +0000 UTC" firstStartedPulling="2025-12-09 10:13:40.809530989 +0000 UTC m=+1577.144035656" lastFinishedPulling="2025-12-09 10:13:54.89779673 +0000 UTC m=+1591.232301397" observedRunningTime="2025-12-09 10:13:55.806262657 +0000 UTC m=+1592.140767324" watchObservedRunningTime="2025-12-09 10:13:55.807877438 +0000 UTC m=+1592.142382125" Dec 09 10:13:56 crc kubenswrapper[4824]: I1209 10:13:56.439824 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="08aaff98-7331-4684-8a51-f009911c20c8" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 09 10:14:04 crc kubenswrapper[4824]: I1209 10:14:04.912050 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:14:04 crc kubenswrapper[4824]: I1209 10:14:04.964630 4824 generic.go:334] "Generic (PLEG): container finished" podID="08aaff98-7331-4684-8a51-f009911c20c8" containerID="5c951d74cbe52d50d259dc7c9f784a070563d7903262cb67cd1ffd8bfa4fa207" exitCode=137 Dec 09 10:14:04 crc kubenswrapper[4824]: I1209 10:14:04.964673 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08aaff98-7331-4684-8a51-f009911c20c8","Type":"ContainerDied","Data":"5c951d74cbe52d50d259dc7c9f784a070563d7903262cb67cd1ffd8bfa4fa207"} Dec 09 10:14:04 crc kubenswrapper[4824]: I1209 10:14:04.964703 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08aaff98-7331-4684-8a51-f009911c20c8","Type":"ContainerDied","Data":"7ffaf0e75bb2ab2cf54c0358c0573b709925d5ca0fcd7d17f50d6db0c15f0c54"} Dec 09 10:14:04 crc kubenswrapper[4824]: I1209 10:14:04.964718 4824 scope.go:117] "RemoveContainer" containerID="5c951d74cbe52d50d259dc7c9f784a070563d7903262cb67cd1ffd8bfa4fa207" Dec 09 10:14:04 crc kubenswrapper[4824]: I1209 10:14:04.964883 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.014765 4824 scope.go:117] "RemoveContainer" containerID="bdf260dcf2266fec4424fb0cfa053c306cc4594001b912d41b973875c1c987e5" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.058702 4824 scope.go:117] "RemoveContainer" containerID="cdba7844f942fd84a6bf5c803d335ce730f79f941f445a120a8193c83026e3bc" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.082155 4824 scope.go:117] "RemoveContainer" containerID="356f5cd708b9ba96569487fc7eaad00284afc25ddc7a67736c6f22e40e529ad4" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.085768 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08aaff98-7331-4684-8a51-f009911c20c8-log-httpd\") pod \"08aaff98-7331-4684-8a51-f009911c20c8\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.085823 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/08aaff98-7331-4684-8a51-f009911c20c8-sg-core-conf-yaml\") pod \"08aaff98-7331-4684-8a51-f009911c20c8\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.085865 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08aaff98-7331-4684-8a51-f009911c20c8-config-data\") pod \"08aaff98-7331-4684-8a51-f009911c20c8\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.085925 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08aaff98-7331-4684-8a51-f009911c20c8-scripts\") pod \"08aaff98-7331-4684-8a51-f009911c20c8\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.086595 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08aaff98-7331-4684-8a51-f009911c20c8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "08aaff98-7331-4684-8a51-f009911c20c8" (UID: "08aaff98-7331-4684-8a51-f009911c20c8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.086841 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wc4vs\" (UniqueName: \"kubernetes.io/projected/08aaff98-7331-4684-8a51-f009911c20c8-kube-api-access-wc4vs\") pod \"08aaff98-7331-4684-8a51-f009911c20c8\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.086991 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08aaff98-7331-4684-8a51-f009911c20c8-run-httpd\") pod \"08aaff98-7331-4684-8a51-f009911c20c8\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.087050 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08aaff98-7331-4684-8a51-f009911c20c8-combined-ca-bundle\") pod \"08aaff98-7331-4684-8a51-f009911c20c8\" (UID: \"08aaff98-7331-4684-8a51-f009911c20c8\") " Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.087919 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08aaff98-7331-4684-8a51-f009911c20c8-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.087945 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08aaff98-7331-4684-8a51-f009911c20c8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "08aaff98-7331-4684-8a51-f009911c20c8" (UID: "08aaff98-7331-4684-8a51-f009911c20c8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.094711 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08aaff98-7331-4684-8a51-f009911c20c8-scripts" (OuterVolumeSpecName: "scripts") pod "08aaff98-7331-4684-8a51-f009911c20c8" (UID: "08aaff98-7331-4684-8a51-f009911c20c8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.107197 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08aaff98-7331-4684-8a51-f009911c20c8-kube-api-access-wc4vs" (OuterVolumeSpecName: "kube-api-access-wc4vs") pod "08aaff98-7331-4684-8a51-f009911c20c8" (UID: "08aaff98-7331-4684-8a51-f009911c20c8"). InnerVolumeSpecName "kube-api-access-wc4vs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.117017 4824 scope.go:117] "RemoveContainer" containerID="5c951d74cbe52d50d259dc7c9f784a070563d7903262cb67cd1ffd8bfa4fa207" Dec 09 10:14:05 crc kubenswrapper[4824]: E1209 10:14:05.117799 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c951d74cbe52d50d259dc7c9f784a070563d7903262cb67cd1ffd8bfa4fa207\": container with ID starting with 5c951d74cbe52d50d259dc7c9f784a070563d7903262cb67cd1ffd8bfa4fa207 not found: ID does not exist" containerID="5c951d74cbe52d50d259dc7c9f784a070563d7903262cb67cd1ffd8bfa4fa207" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.117873 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c951d74cbe52d50d259dc7c9f784a070563d7903262cb67cd1ffd8bfa4fa207"} err="failed to get container status \"5c951d74cbe52d50d259dc7c9f784a070563d7903262cb67cd1ffd8bfa4fa207\": rpc error: code = NotFound desc = could not find container \"5c951d74cbe52d50d259dc7c9f784a070563d7903262cb67cd1ffd8bfa4fa207\": container with ID starting with 5c951d74cbe52d50d259dc7c9f784a070563d7903262cb67cd1ffd8bfa4fa207 not found: ID does not exist" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.117911 4824 scope.go:117] "RemoveContainer" containerID="bdf260dcf2266fec4424fb0cfa053c306cc4594001b912d41b973875c1c987e5" Dec 09 10:14:05 crc kubenswrapper[4824]: E1209 10:14:05.119237 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdf260dcf2266fec4424fb0cfa053c306cc4594001b912d41b973875c1c987e5\": container with ID starting with bdf260dcf2266fec4424fb0cfa053c306cc4594001b912d41b973875c1c987e5 not found: ID does not exist" containerID="bdf260dcf2266fec4424fb0cfa053c306cc4594001b912d41b973875c1c987e5" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.119261 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdf260dcf2266fec4424fb0cfa053c306cc4594001b912d41b973875c1c987e5"} err="failed to get container status \"bdf260dcf2266fec4424fb0cfa053c306cc4594001b912d41b973875c1c987e5\": rpc error: code = NotFound desc = could not find container \"bdf260dcf2266fec4424fb0cfa053c306cc4594001b912d41b973875c1c987e5\": container with ID starting with bdf260dcf2266fec4424fb0cfa053c306cc4594001b912d41b973875c1c987e5 not found: ID does not exist" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.119280 4824 scope.go:117] "RemoveContainer" containerID="cdba7844f942fd84a6bf5c803d335ce730f79f941f445a120a8193c83026e3bc" Dec 09 10:14:05 crc kubenswrapper[4824]: E1209 10:14:05.119856 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdba7844f942fd84a6bf5c803d335ce730f79f941f445a120a8193c83026e3bc\": container with ID starting with cdba7844f942fd84a6bf5c803d335ce730f79f941f445a120a8193c83026e3bc not found: ID does not exist" containerID="cdba7844f942fd84a6bf5c803d335ce730f79f941f445a120a8193c83026e3bc" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.119939 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdba7844f942fd84a6bf5c803d335ce730f79f941f445a120a8193c83026e3bc"} err="failed to get container status \"cdba7844f942fd84a6bf5c803d335ce730f79f941f445a120a8193c83026e3bc\": rpc error: code = NotFound desc = could not find container \"cdba7844f942fd84a6bf5c803d335ce730f79f941f445a120a8193c83026e3bc\": container with ID starting with cdba7844f942fd84a6bf5c803d335ce730f79f941f445a120a8193c83026e3bc not found: ID does not exist" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.120023 4824 scope.go:117] "RemoveContainer" containerID="356f5cd708b9ba96569487fc7eaad00284afc25ddc7a67736c6f22e40e529ad4" Dec 09 10:14:05 crc kubenswrapper[4824]: E1209 10:14:05.120710 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"356f5cd708b9ba96569487fc7eaad00284afc25ddc7a67736c6f22e40e529ad4\": container with ID starting with 356f5cd708b9ba96569487fc7eaad00284afc25ddc7a67736c6f22e40e529ad4 not found: ID does not exist" containerID="356f5cd708b9ba96569487fc7eaad00284afc25ddc7a67736c6f22e40e529ad4" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.120751 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"356f5cd708b9ba96569487fc7eaad00284afc25ddc7a67736c6f22e40e529ad4"} err="failed to get container status \"356f5cd708b9ba96569487fc7eaad00284afc25ddc7a67736c6f22e40e529ad4\": rpc error: code = NotFound desc = could not find container \"356f5cd708b9ba96569487fc7eaad00284afc25ddc7a67736c6f22e40e529ad4\": container with ID starting with 356f5cd708b9ba96569487fc7eaad00284afc25ddc7a67736c6f22e40e529ad4 not found: ID does not exist" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.126918 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08aaff98-7331-4684-8a51-f009911c20c8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "08aaff98-7331-4684-8a51-f009911c20c8" (UID: "08aaff98-7331-4684-8a51-f009911c20c8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.311124 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08aaff98-7331-4684-8a51-f009911c20c8-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.311165 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/08aaff98-7331-4684-8a51-f009911c20c8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.311183 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08aaff98-7331-4684-8a51-f009911c20c8-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.311196 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wc4vs\" (UniqueName: \"kubernetes.io/projected/08aaff98-7331-4684-8a51-f009911c20c8-kube-api-access-wc4vs\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.360921 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08aaff98-7331-4684-8a51-f009911c20c8-config-data" (OuterVolumeSpecName: "config-data") pod "08aaff98-7331-4684-8a51-f009911c20c8" (UID: "08aaff98-7331-4684-8a51-f009911c20c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.371549 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08aaff98-7331-4684-8a51-f009911c20c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "08aaff98-7331-4684-8a51-f009911c20c8" (UID: "08aaff98-7331-4684-8a51-f009911c20c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.416651 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08aaff98-7331-4684-8a51-f009911c20c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.416697 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08aaff98-7331-4684-8a51-f009911c20c8-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.604247 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.618630 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.645351 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:14:05 crc kubenswrapper[4824]: E1209 10:14:05.646018 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04bcf8d0-9439-4739-a503-e504651a80dd" containerName="heat-engine" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.646045 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="04bcf8d0-9439-4739-a503-e504651a80dd" containerName="heat-engine" Dec 09 10:14:05 crc kubenswrapper[4824]: E1209 10:14:05.646071 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08aaff98-7331-4684-8a51-f009911c20c8" containerName="ceilometer-central-agent" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.646081 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="08aaff98-7331-4684-8a51-f009911c20c8" containerName="ceilometer-central-agent" Dec 09 10:14:05 crc kubenswrapper[4824]: E1209 10:14:05.646091 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08aaff98-7331-4684-8a51-f009911c20c8" containerName="sg-core" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.646099 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="08aaff98-7331-4684-8a51-f009911c20c8" containerName="sg-core" Dec 09 10:14:05 crc kubenswrapper[4824]: E1209 10:14:05.646113 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08aaff98-7331-4684-8a51-f009911c20c8" containerName="ceilometer-notification-agent" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.646124 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="08aaff98-7331-4684-8a51-f009911c20c8" containerName="ceilometer-notification-agent" Dec 09 10:14:05 crc kubenswrapper[4824]: E1209 10:14:05.646151 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08aaff98-7331-4684-8a51-f009911c20c8" containerName="proxy-httpd" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.646158 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="08aaff98-7331-4684-8a51-f009911c20c8" containerName="proxy-httpd" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.646486 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="08aaff98-7331-4684-8a51-f009911c20c8" containerName="sg-core" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.646517 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="04bcf8d0-9439-4739-a503-e504651a80dd" containerName="heat-engine" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.646537 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="08aaff98-7331-4684-8a51-f009911c20c8" containerName="ceilometer-notification-agent" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.646550 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="08aaff98-7331-4684-8a51-f009911c20c8" containerName="proxy-httpd" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.646579 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="08aaff98-7331-4684-8a51-f009911c20c8" containerName="ceilometer-central-agent" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.649537 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.652703 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.654287 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.658718 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.723924 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5244f0d6-a488-4476-af14-85300dd6226b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " pod="openstack/ceilometer-0" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.724378 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5244f0d6-a488-4476-af14-85300dd6226b-config-data\") pod \"ceilometer-0\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " pod="openstack/ceilometer-0" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.724526 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5244f0d6-a488-4476-af14-85300dd6226b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " pod="openstack/ceilometer-0" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.724592 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5244f0d6-a488-4476-af14-85300dd6226b-log-httpd\") pod \"ceilometer-0\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " pod="openstack/ceilometer-0" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.724747 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5244f0d6-a488-4476-af14-85300dd6226b-run-httpd\") pod \"ceilometer-0\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " pod="openstack/ceilometer-0" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.724842 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5244f0d6-a488-4476-af14-85300dd6226b-scripts\") pod \"ceilometer-0\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " pod="openstack/ceilometer-0" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.724876 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pclsr\" (UniqueName: \"kubernetes.io/projected/5244f0d6-a488-4476-af14-85300dd6226b-kube-api-access-pclsr\") pod \"ceilometer-0\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " pod="openstack/ceilometer-0" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.827268 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5244f0d6-a488-4476-af14-85300dd6226b-log-httpd\") pod \"ceilometer-0\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " pod="openstack/ceilometer-0" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.827415 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5244f0d6-a488-4476-af14-85300dd6226b-run-httpd\") pod \"ceilometer-0\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " pod="openstack/ceilometer-0" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.827474 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5244f0d6-a488-4476-af14-85300dd6226b-scripts\") pod \"ceilometer-0\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " pod="openstack/ceilometer-0" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.827501 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pclsr\" (UniqueName: \"kubernetes.io/projected/5244f0d6-a488-4476-af14-85300dd6226b-kube-api-access-pclsr\") pod \"ceilometer-0\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " pod="openstack/ceilometer-0" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.827565 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5244f0d6-a488-4476-af14-85300dd6226b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " pod="openstack/ceilometer-0" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.827594 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5244f0d6-a488-4476-af14-85300dd6226b-config-data\") pod \"ceilometer-0\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " pod="openstack/ceilometer-0" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.827665 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5244f0d6-a488-4476-af14-85300dd6226b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " pod="openstack/ceilometer-0" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.827903 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5244f0d6-a488-4476-af14-85300dd6226b-log-httpd\") pod \"ceilometer-0\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " pod="openstack/ceilometer-0" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.828079 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5244f0d6-a488-4476-af14-85300dd6226b-run-httpd\") pod \"ceilometer-0\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " pod="openstack/ceilometer-0" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.835398 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5244f0d6-a488-4476-af14-85300dd6226b-config-data\") pod \"ceilometer-0\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " pod="openstack/ceilometer-0" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.835675 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5244f0d6-a488-4476-af14-85300dd6226b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " pod="openstack/ceilometer-0" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.836129 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5244f0d6-a488-4476-af14-85300dd6226b-scripts\") pod \"ceilometer-0\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " pod="openstack/ceilometer-0" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.841013 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5244f0d6-a488-4476-af14-85300dd6226b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " pod="openstack/ceilometer-0" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.848592 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pclsr\" (UniqueName: \"kubernetes.io/projected/5244f0d6-a488-4476-af14-85300dd6226b-kube-api-access-pclsr\") pod \"ceilometer-0\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " pod="openstack/ceilometer-0" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.937139 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08aaff98-7331-4684-8a51-f009911c20c8" path="/var/lib/kubelet/pods/08aaff98-7331-4684-8a51-f009911c20c8/volumes" Dec 09 10:14:05 crc kubenswrapper[4824]: I1209 10:14:05.969614 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:14:06 crc kubenswrapper[4824]: I1209 10:14:06.511355 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:14:07 crc kubenswrapper[4824]: I1209 10:14:07.015587 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5244f0d6-a488-4476-af14-85300dd6226b","Type":"ContainerStarted","Data":"e9dea85f6258bfb6e908c24472c6800628fe11fba80051094c1e11dea63e2cca"} Dec 09 10:14:08 crc kubenswrapper[4824]: I1209 10:14:08.156493 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5244f0d6-a488-4476-af14-85300dd6226b","Type":"ContainerStarted","Data":"9577b6932c99285915388271359aafef834851ccde28f913db06e54257004bb2"} Dec 09 10:14:09 crc kubenswrapper[4824]: I1209 10:14:09.173296 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5244f0d6-a488-4476-af14-85300dd6226b","Type":"ContainerStarted","Data":"1534eba33de7b8c69c5492f42327e161d47d5d79d14bd27c19398383b60b7f8c"} Dec 09 10:14:10 crc kubenswrapper[4824]: I1209 10:14:10.201170 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5244f0d6-a488-4476-af14-85300dd6226b","Type":"ContainerStarted","Data":"170f6063b7481c1ef8bd912e1515224b8d9ae229f66c58cd7305b7ad5de4e971"} Dec 09 10:14:10 crc kubenswrapper[4824]: I1209 10:14:10.210050 4824 generic.go:334] "Generic (PLEG): container finished" podID="221367da-a9dd-488d-a2c5-69bc0f57e589" containerID="70b7050442fdf6eb10381f437e45fa00761a5e37ce90432c54bcefc8cd9db7b2" exitCode=0 Dec 09 10:14:10 crc kubenswrapper[4824]: I1209 10:14:10.210108 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8q4b9" event={"ID":"221367da-a9dd-488d-a2c5-69bc0f57e589","Type":"ContainerDied","Data":"70b7050442fdf6eb10381f437e45fa00761a5e37ce90432c54bcefc8cd9db7b2"} Dec 09 10:14:11 crc kubenswrapper[4824]: I1209 10:14:11.499728 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5244f0d6-a488-4476-af14-85300dd6226b","Type":"ContainerStarted","Data":"bc92c693e73c1ec6e644fb694076152511dfc209f591be3d4bef458b8a909d48"} Dec 09 10:14:11 crc kubenswrapper[4824]: I1209 10:14:11.500357 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 10:14:11 crc kubenswrapper[4824]: I1209 10:14:11.538636 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.352534737 podStartE2EDuration="6.538591617s" podCreationTimestamp="2025-12-09 10:14:05 +0000 UTC" firstStartedPulling="2025-12-09 10:14:06.515422962 +0000 UTC m=+1602.849927619" lastFinishedPulling="2025-12-09 10:14:10.701479832 +0000 UTC m=+1607.035984499" observedRunningTime="2025-12-09 10:14:11.524193239 +0000 UTC m=+1607.858697906" watchObservedRunningTime="2025-12-09 10:14:11.538591617 +0000 UTC m=+1607.873096284" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.058816 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8q4b9" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.248339 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/221367da-a9dd-488d-a2c5-69bc0f57e589-scripts\") pod \"221367da-a9dd-488d-a2c5-69bc0f57e589\" (UID: \"221367da-a9dd-488d-a2c5-69bc0f57e589\") " Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.248490 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngt5h\" (UniqueName: \"kubernetes.io/projected/221367da-a9dd-488d-a2c5-69bc0f57e589-kube-api-access-ngt5h\") pod \"221367da-a9dd-488d-a2c5-69bc0f57e589\" (UID: \"221367da-a9dd-488d-a2c5-69bc0f57e589\") " Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.249637 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/221367da-a9dd-488d-a2c5-69bc0f57e589-config-data\") pod \"221367da-a9dd-488d-a2c5-69bc0f57e589\" (UID: \"221367da-a9dd-488d-a2c5-69bc0f57e589\") " Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.249732 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/221367da-a9dd-488d-a2c5-69bc0f57e589-combined-ca-bundle\") pod \"221367da-a9dd-488d-a2c5-69bc0f57e589\" (UID: \"221367da-a9dd-488d-a2c5-69bc0f57e589\") " Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.255195 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/221367da-a9dd-488d-a2c5-69bc0f57e589-scripts" (OuterVolumeSpecName: "scripts") pod "221367da-a9dd-488d-a2c5-69bc0f57e589" (UID: "221367da-a9dd-488d-a2c5-69bc0f57e589"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.255969 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/221367da-a9dd-488d-a2c5-69bc0f57e589-kube-api-access-ngt5h" (OuterVolumeSpecName: "kube-api-access-ngt5h") pod "221367da-a9dd-488d-a2c5-69bc0f57e589" (UID: "221367da-a9dd-488d-a2c5-69bc0f57e589"). InnerVolumeSpecName "kube-api-access-ngt5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.300821 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/221367da-a9dd-488d-a2c5-69bc0f57e589-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "221367da-a9dd-488d-a2c5-69bc0f57e589" (UID: "221367da-a9dd-488d-a2c5-69bc0f57e589"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.301535 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/221367da-a9dd-488d-a2c5-69bc0f57e589-config-data" (OuterVolumeSpecName: "config-data") pod "221367da-a9dd-488d-a2c5-69bc0f57e589" (UID: "221367da-a9dd-488d-a2c5-69bc0f57e589"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.352569 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngt5h\" (UniqueName: \"kubernetes.io/projected/221367da-a9dd-488d-a2c5-69bc0f57e589-kube-api-access-ngt5h\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.352615 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/221367da-a9dd-488d-a2c5-69bc0f57e589-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.352634 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/221367da-a9dd-488d-a2c5-69bc0f57e589-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.352644 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/221367da-a9dd-488d-a2c5-69bc0f57e589-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.511581 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8q4b9" event={"ID":"221367da-a9dd-488d-a2c5-69bc0f57e589","Type":"ContainerDied","Data":"784eac1cecb8dce7d30876abcd3edcdfb80bd51c6d28ba85a0238a0690e39a59"} Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.511663 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="784eac1cecb8dce7d30876abcd3edcdfb80bd51c6d28ba85a0238a0690e39a59" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.511622 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8q4b9" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.561410 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 09 10:14:12 crc kubenswrapper[4824]: E1209 10:14:12.562291 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="221367da-a9dd-488d-a2c5-69bc0f57e589" containerName="nova-cell0-conductor-db-sync" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.562378 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="221367da-a9dd-488d-a2c5-69bc0f57e589" containerName="nova-cell0-conductor-db-sync" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.562708 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="221367da-a9dd-488d-a2c5-69bc0f57e589" containerName="nova-cell0-conductor-db-sync" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.563679 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.566638 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-8t2cn" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.567001 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.580886 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.761129 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97ltr\" (UniqueName: \"kubernetes.io/projected/71b5e2f1-303e-4718-be1f-fb2ab839dd64-kube-api-access-97ltr\") pod \"nova-cell0-conductor-0\" (UID: \"71b5e2f1-303e-4718-be1f-fb2ab839dd64\") " pod="openstack/nova-cell0-conductor-0" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.761236 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71b5e2f1-303e-4718-be1f-fb2ab839dd64-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"71b5e2f1-303e-4718-be1f-fb2ab839dd64\") " pod="openstack/nova-cell0-conductor-0" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.761325 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b5e2f1-303e-4718-be1f-fb2ab839dd64-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"71b5e2f1-303e-4718-be1f-fb2ab839dd64\") " pod="openstack/nova-cell0-conductor-0" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.864036 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97ltr\" (UniqueName: \"kubernetes.io/projected/71b5e2f1-303e-4718-be1f-fb2ab839dd64-kube-api-access-97ltr\") pod \"nova-cell0-conductor-0\" (UID: \"71b5e2f1-303e-4718-be1f-fb2ab839dd64\") " pod="openstack/nova-cell0-conductor-0" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.864193 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71b5e2f1-303e-4718-be1f-fb2ab839dd64-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"71b5e2f1-303e-4718-be1f-fb2ab839dd64\") " pod="openstack/nova-cell0-conductor-0" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.864322 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b5e2f1-303e-4718-be1f-fb2ab839dd64-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"71b5e2f1-303e-4718-be1f-fb2ab839dd64\") " pod="openstack/nova-cell0-conductor-0" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.940274 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b5e2f1-303e-4718-be1f-fb2ab839dd64-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"71b5e2f1-303e-4718-be1f-fb2ab839dd64\") " pod="openstack/nova-cell0-conductor-0" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.940292 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71b5e2f1-303e-4718-be1f-fb2ab839dd64-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"71b5e2f1-303e-4718-be1f-fb2ab839dd64\") " pod="openstack/nova-cell0-conductor-0" Dec 09 10:14:12 crc kubenswrapper[4824]: I1209 10:14:12.953423 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97ltr\" (UniqueName: \"kubernetes.io/projected/71b5e2f1-303e-4718-be1f-fb2ab839dd64-kube-api-access-97ltr\") pod \"nova-cell0-conductor-0\" (UID: \"71b5e2f1-303e-4718-be1f-fb2ab839dd64\") " pod="openstack/nova-cell0-conductor-0" Dec 09 10:14:13 crc kubenswrapper[4824]: I1209 10:14:13.185248 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 09 10:14:13 crc kubenswrapper[4824]: I1209 10:14:13.778825 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:14:13 crc kubenswrapper[4824]: I1209 10:14:13.779972 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5244f0d6-a488-4476-af14-85300dd6226b" containerName="ceilometer-central-agent" containerID="cri-o://9577b6932c99285915388271359aafef834851ccde28f913db06e54257004bb2" gracePeriod=30 Dec 09 10:14:13 crc kubenswrapper[4824]: I1209 10:14:13.780108 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5244f0d6-a488-4476-af14-85300dd6226b" containerName="sg-core" containerID="cri-o://170f6063b7481c1ef8bd912e1515224b8d9ae229f66c58cd7305b7ad5de4e971" gracePeriod=30 Dec 09 10:14:13 crc kubenswrapper[4824]: I1209 10:14:13.780126 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5244f0d6-a488-4476-af14-85300dd6226b" containerName="proxy-httpd" containerID="cri-o://bc92c693e73c1ec6e644fb694076152511dfc209f591be3d4bef458b8a909d48" gracePeriod=30 Dec 09 10:14:13 crc kubenswrapper[4824]: I1209 10:14:13.780665 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5244f0d6-a488-4476-af14-85300dd6226b" containerName="ceilometer-notification-agent" containerID="cri-o://1534eba33de7b8c69c5492f42327e161d47d5d79d14bd27c19398383b60b7f8c" gracePeriod=30 Dec 09 10:14:14 crc kubenswrapper[4824]: I1209 10:14:14.082351 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 09 10:14:14 crc kubenswrapper[4824]: E1209 10:14:14.444015 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5244f0d6_a488_4476_af14_85300dd6226b.slice/crio-bc92c693e73c1ec6e644fb694076152511dfc209f591be3d4bef458b8a909d48.scope\": RecentStats: unable to find data in memory cache]" Dec 09 10:14:14 crc kubenswrapper[4824]: I1209 10:14:14.557892 4824 generic.go:334] "Generic (PLEG): container finished" podID="5244f0d6-a488-4476-af14-85300dd6226b" containerID="bc92c693e73c1ec6e644fb694076152511dfc209f591be3d4bef458b8a909d48" exitCode=0 Dec 09 10:14:14 crc kubenswrapper[4824]: I1209 10:14:14.557926 4824 generic.go:334] "Generic (PLEG): container finished" podID="5244f0d6-a488-4476-af14-85300dd6226b" containerID="170f6063b7481c1ef8bd912e1515224b8d9ae229f66c58cd7305b7ad5de4e971" exitCode=2 Dec 09 10:14:14 crc kubenswrapper[4824]: I1209 10:14:14.558008 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5244f0d6-a488-4476-af14-85300dd6226b","Type":"ContainerDied","Data":"bc92c693e73c1ec6e644fb694076152511dfc209f591be3d4bef458b8a909d48"} Dec 09 10:14:14 crc kubenswrapper[4824]: I1209 10:14:14.558146 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5244f0d6-a488-4476-af14-85300dd6226b","Type":"ContainerDied","Data":"170f6063b7481c1ef8bd912e1515224b8d9ae229f66c58cd7305b7ad5de4e971"} Dec 09 10:14:14 crc kubenswrapper[4824]: I1209 10:14:14.563370 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"71b5e2f1-303e-4718-be1f-fb2ab839dd64","Type":"ContainerStarted","Data":"8227f6d3c1b590ad19d3011aaeb987bdf0511fb1d5e867cf1d9c29d45ccc9dfa"} Dec 09 10:14:15 crc kubenswrapper[4824]: I1209 10:14:15.684315 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"71b5e2f1-303e-4718-be1f-fb2ab839dd64","Type":"ContainerStarted","Data":"199c752a2426bc2b3dce59e49a86d58691474befa1f3d15377847c12fa5c394c"} Dec 09 10:14:15 crc kubenswrapper[4824]: I1209 10:14:15.685440 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 09 10:14:15 crc kubenswrapper[4824]: I1209 10:14:15.689827 4824 generic.go:334] "Generic (PLEG): container finished" podID="5244f0d6-a488-4476-af14-85300dd6226b" containerID="1534eba33de7b8c69c5492f42327e161d47d5d79d14bd27c19398383b60b7f8c" exitCode=0 Dec 09 10:14:15 crc kubenswrapper[4824]: I1209 10:14:15.689867 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5244f0d6-a488-4476-af14-85300dd6226b","Type":"ContainerDied","Data":"1534eba33de7b8c69c5492f42327e161d47d5d79d14bd27c19398383b60b7f8c"} Dec 09 10:14:15 crc kubenswrapper[4824]: I1209 10:14:15.720687 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=3.720659919 podStartE2EDuration="3.720659919s" podCreationTimestamp="2025-12-09 10:14:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:14:15.711246501 +0000 UTC m=+1612.045751178" watchObservedRunningTime="2025-12-09 10:14:15.720659919 +0000 UTC m=+1612.055164586" Dec 09 10:14:21 crc kubenswrapper[4824]: I1209 10:14:21.794354 4824 generic.go:334] "Generic (PLEG): container finished" podID="5244f0d6-a488-4476-af14-85300dd6226b" containerID="9577b6932c99285915388271359aafef834851ccde28f913db06e54257004bb2" exitCode=0 Dec 09 10:14:21 crc kubenswrapper[4824]: I1209 10:14:21.795087 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5244f0d6-a488-4476-af14-85300dd6226b","Type":"ContainerDied","Data":"9577b6932c99285915388271359aafef834851ccde28f913db06e54257004bb2"} Dec 09 10:14:21 crc kubenswrapper[4824]: I1209 10:14:21.974088 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:14:21 crc kubenswrapper[4824]: I1209 10:14:21.984762 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5244f0d6-a488-4476-af14-85300dd6226b-run-httpd\") pod \"5244f0d6-a488-4476-af14-85300dd6226b\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " Dec 09 10:14:21 crc kubenswrapper[4824]: I1209 10:14:21.986363 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5244f0d6-a488-4476-af14-85300dd6226b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5244f0d6-a488-4476-af14-85300dd6226b" (UID: "5244f0d6-a488-4476-af14-85300dd6226b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.089529 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5244f0d6-a488-4476-af14-85300dd6226b-scripts\") pod \"5244f0d6-a488-4476-af14-85300dd6226b\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.089641 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5244f0d6-a488-4476-af14-85300dd6226b-sg-core-conf-yaml\") pod \"5244f0d6-a488-4476-af14-85300dd6226b\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.089697 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5244f0d6-a488-4476-af14-85300dd6226b-combined-ca-bundle\") pod \"5244f0d6-a488-4476-af14-85300dd6226b\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.089736 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5244f0d6-a488-4476-af14-85300dd6226b-config-data\") pod \"5244f0d6-a488-4476-af14-85300dd6226b\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.089823 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5244f0d6-a488-4476-af14-85300dd6226b-log-httpd\") pod \"5244f0d6-a488-4476-af14-85300dd6226b\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.089951 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pclsr\" (UniqueName: \"kubernetes.io/projected/5244f0d6-a488-4476-af14-85300dd6226b-kube-api-access-pclsr\") pod \"5244f0d6-a488-4476-af14-85300dd6226b\" (UID: \"5244f0d6-a488-4476-af14-85300dd6226b\") " Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.094104 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5244f0d6-a488-4476-af14-85300dd6226b-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.094542 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5244f0d6-a488-4476-af14-85300dd6226b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5244f0d6-a488-4476-af14-85300dd6226b" (UID: "5244f0d6-a488-4476-af14-85300dd6226b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.098897 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5244f0d6-a488-4476-af14-85300dd6226b-kube-api-access-pclsr" (OuterVolumeSpecName: "kube-api-access-pclsr") pod "5244f0d6-a488-4476-af14-85300dd6226b" (UID: "5244f0d6-a488-4476-af14-85300dd6226b"). InnerVolumeSpecName "kube-api-access-pclsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.102690 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5244f0d6-a488-4476-af14-85300dd6226b-scripts" (OuterVolumeSpecName: "scripts") pod "5244f0d6-a488-4476-af14-85300dd6226b" (UID: "5244f0d6-a488-4476-af14-85300dd6226b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.105347 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-52bnj"] Dec 09 10:14:22 crc kubenswrapper[4824]: E1209 10:14:22.109545 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5244f0d6-a488-4476-af14-85300dd6226b" containerName="ceilometer-notification-agent" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.109575 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5244f0d6-a488-4476-af14-85300dd6226b" containerName="ceilometer-notification-agent" Dec 09 10:14:22 crc kubenswrapper[4824]: E1209 10:14:22.109604 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5244f0d6-a488-4476-af14-85300dd6226b" containerName="sg-core" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.109614 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5244f0d6-a488-4476-af14-85300dd6226b" containerName="sg-core" Dec 09 10:14:22 crc kubenswrapper[4824]: E1209 10:14:22.109630 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5244f0d6-a488-4476-af14-85300dd6226b" containerName="proxy-httpd" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.109639 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5244f0d6-a488-4476-af14-85300dd6226b" containerName="proxy-httpd" Dec 09 10:14:22 crc kubenswrapper[4824]: E1209 10:14:22.109660 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5244f0d6-a488-4476-af14-85300dd6226b" containerName="ceilometer-central-agent" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.109666 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5244f0d6-a488-4476-af14-85300dd6226b" containerName="ceilometer-central-agent" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.109968 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5244f0d6-a488-4476-af14-85300dd6226b" containerName="ceilometer-notification-agent" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.109987 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5244f0d6-a488-4476-af14-85300dd6226b" containerName="sg-core" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.110012 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5244f0d6-a488-4476-af14-85300dd6226b" containerName="proxy-httpd" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.110030 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5244f0d6-a488-4476-af14-85300dd6226b" containerName="ceilometer-central-agent" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.111519 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-52bnj" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.147192 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5244f0d6-a488-4476-af14-85300dd6226b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5244f0d6-a488-4476-af14-85300dd6226b" (UID: "5244f0d6-a488-4476-af14-85300dd6226b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.152481 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-f925-account-create-update-zk59p"] Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.154750 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-f925-account-create-update-zk59p" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.159775 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.195686 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-52bnj"] Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.199751 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wn7n\" (UniqueName: \"kubernetes.io/projected/3727faad-272d-4a6a-a2ba-7b5be05cd11b-kube-api-access-4wn7n\") pod \"aodh-db-create-52bnj\" (UID: \"3727faad-272d-4a6a-a2ba-7b5be05cd11b\") " pod="openstack/aodh-db-create-52bnj" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.200065 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgkx6\" (UniqueName: \"kubernetes.io/projected/96ede01d-246b-47e6-ac4d-06c74436006c-kube-api-access-fgkx6\") pod \"aodh-f925-account-create-update-zk59p\" (UID: \"96ede01d-246b-47e6-ac4d-06c74436006c\") " pod="openstack/aodh-f925-account-create-update-zk59p" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.200466 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3727faad-272d-4a6a-a2ba-7b5be05cd11b-operator-scripts\") pod \"aodh-db-create-52bnj\" (UID: \"3727faad-272d-4a6a-a2ba-7b5be05cd11b\") " pod="openstack/aodh-db-create-52bnj" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.200561 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96ede01d-246b-47e6-ac4d-06c74436006c-operator-scripts\") pod \"aodh-f925-account-create-update-zk59p\" (UID: \"96ede01d-246b-47e6-ac4d-06c74436006c\") " pod="openstack/aodh-f925-account-create-update-zk59p" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.200959 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5244f0d6-a488-4476-af14-85300dd6226b-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.200987 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5244f0d6-a488-4476-af14-85300dd6226b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.200999 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5244f0d6-a488-4476-af14-85300dd6226b-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.201014 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pclsr\" (UniqueName: \"kubernetes.io/projected/5244f0d6-a488-4476-af14-85300dd6226b-kube-api-access-pclsr\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.216947 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-f925-account-create-update-zk59p"] Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.252748 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5244f0d6-a488-4476-af14-85300dd6226b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5244f0d6-a488-4476-af14-85300dd6226b" (UID: "5244f0d6-a488-4476-af14-85300dd6226b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.302630 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgkx6\" (UniqueName: \"kubernetes.io/projected/96ede01d-246b-47e6-ac4d-06c74436006c-kube-api-access-fgkx6\") pod \"aodh-f925-account-create-update-zk59p\" (UID: \"96ede01d-246b-47e6-ac4d-06c74436006c\") " pod="openstack/aodh-f925-account-create-update-zk59p" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.302822 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3727faad-272d-4a6a-a2ba-7b5be05cd11b-operator-scripts\") pod \"aodh-db-create-52bnj\" (UID: \"3727faad-272d-4a6a-a2ba-7b5be05cd11b\") " pod="openstack/aodh-db-create-52bnj" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.302868 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96ede01d-246b-47e6-ac4d-06c74436006c-operator-scripts\") pod \"aodh-f925-account-create-update-zk59p\" (UID: \"96ede01d-246b-47e6-ac4d-06c74436006c\") " pod="openstack/aodh-f925-account-create-update-zk59p" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.302986 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wn7n\" (UniqueName: \"kubernetes.io/projected/3727faad-272d-4a6a-a2ba-7b5be05cd11b-kube-api-access-4wn7n\") pod \"aodh-db-create-52bnj\" (UID: \"3727faad-272d-4a6a-a2ba-7b5be05cd11b\") " pod="openstack/aodh-db-create-52bnj" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.303102 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5244f0d6-a488-4476-af14-85300dd6226b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.304751 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3727faad-272d-4a6a-a2ba-7b5be05cd11b-operator-scripts\") pod \"aodh-db-create-52bnj\" (UID: \"3727faad-272d-4a6a-a2ba-7b5be05cd11b\") " pod="openstack/aodh-db-create-52bnj" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.305477 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96ede01d-246b-47e6-ac4d-06c74436006c-operator-scripts\") pod \"aodh-f925-account-create-update-zk59p\" (UID: \"96ede01d-246b-47e6-ac4d-06c74436006c\") " pod="openstack/aodh-f925-account-create-update-zk59p" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.315647 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5244f0d6-a488-4476-af14-85300dd6226b-config-data" (OuterVolumeSpecName: "config-data") pod "5244f0d6-a488-4476-af14-85300dd6226b" (UID: "5244f0d6-a488-4476-af14-85300dd6226b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.320019 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wn7n\" (UniqueName: \"kubernetes.io/projected/3727faad-272d-4a6a-a2ba-7b5be05cd11b-kube-api-access-4wn7n\") pod \"aodh-db-create-52bnj\" (UID: \"3727faad-272d-4a6a-a2ba-7b5be05cd11b\") " pod="openstack/aodh-db-create-52bnj" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.320459 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgkx6\" (UniqueName: \"kubernetes.io/projected/96ede01d-246b-47e6-ac4d-06c74436006c-kube-api-access-fgkx6\") pod \"aodh-f925-account-create-update-zk59p\" (UID: \"96ede01d-246b-47e6-ac4d-06c74436006c\") " pod="openstack/aodh-f925-account-create-update-zk59p" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.482742 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5244f0d6-a488-4476-af14-85300dd6226b-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.559983 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-52bnj" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.570683 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-f925-account-create-update-zk59p" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.832951 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5244f0d6-a488-4476-af14-85300dd6226b","Type":"ContainerDied","Data":"e9dea85f6258bfb6e908c24472c6800628fe11fba80051094c1e11dea63e2cca"} Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.833201 4824 scope.go:117] "RemoveContainer" containerID="bc92c693e73c1ec6e644fb694076152511dfc209f591be3d4bef458b8a909d48" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.833437 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.920905 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.949140 4824 scope.go:117] "RemoveContainer" containerID="170f6063b7481c1ef8bd912e1515224b8d9ae229f66c58cd7305b7ad5de4e971" Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.952837 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:14:22 crc kubenswrapper[4824]: I1209 10:14:22.976524 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.003529 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.003645 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.009524 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.009532 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.023858 4824 scope.go:117] "RemoveContainer" containerID="1534eba33de7b8c69c5492f42327e161d47d5d79d14bd27c19398383b60b7f8c" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.066528 4824 scope.go:117] "RemoveContainer" containerID="9577b6932c99285915388271359aafef834851ccde28f913db06e54257004bb2" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.105847 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-scripts\") pod \"ceilometer-0\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " pod="openstack/ceilometer-0" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.105975 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-log-httpd\") pod \"ceilometer-0\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " pod="openstack/ceilometer-0" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.106032 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwqh5\" (UniqueName: \"kubernetes.io/projected/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-kube-api-access-mwqh5\") pod \"ceilometer-0\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " pod="openstack/ceilometer-0" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.106168 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-run-httpd\") pod \"ceilometer-0\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " pod="openstack/ceilometer-0" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.106255 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " pod="openstack/ceilometer-0" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.106343 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-config-data\") pod \"ceilometer-0\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " pod="openstack/ceilometer-0" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.106696 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " pod="openstack/ceilometer-0" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.364908 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.413988 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-run-httpd\") pod \"ceilometer-0\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " pod="openstack/ceilometer-0" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.414079 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " pod="openstack/ceilometer-0" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.414130 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-config-data\") pod \"ceilometer-0\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " pod="openstack/ceilometer-0" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.414240 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " pod="openstack/ceilometer-0" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.414306 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-scripts\") pod \"ceilometer-0\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " pod="openstack/ceilometer-0" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.414345 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-log-httpd\") pod \"ceilometer-0\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " pod="openstack/ceilometer-0" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.414365 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwqh5\" (UniqueName: \"kubernetes.io/projected/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-kube-api-access-mwqh5\") pod \"ceilometer-0\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " pod="openstack/ceilometer-0" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.415433 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-run-httpd\") pod \"ceilometer-0\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " pod="openstack/ceilometer-0" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.417196 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-52bnj"] Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.418478 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-log-httpd\") pod \"ceilometer-0\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " pod="openstack/ceilometer-0" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.426700 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-config-data\") pod \"ceilometer-0\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " pod="openstack/ceilometer-0" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.431595 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-scripts\") pod \"ceilometer-0\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " pod="openstack/ceilometer-0" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.432746 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " pod="openstack/ceilometer-0" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.434800 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " pod="openstack/ceilometer-0" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.457270 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwqh5\" (UniqueName: \"kubernetes.io/projected/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-kube-api-access-mwqh5\") pod \"ceilometer-0\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " pod="openstack/ceilometer-0" Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.644262 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-f925-account-create-update-zk59p"] Dec 09 10:14:23 crc kubenswrapper[4824]: I1209 10:14:23.664609 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.011272 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5244f0d6-a488-4476-af14-85300dd6226b" path="/var/lib/kubelet/pods/5244f0d6-a488-4476-af14-85300dd6226b/volumes" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.016747 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-f925-account-create-update-zk59p" event={"ID":"96ede01d-246b-47e6-ac4d-06c74436006c","Type":"ContainerStarted","Data":"c7e00fbabf965185f45930f286488d6096d87daf00d7fff9406b47db876962c8"} Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.020670 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-52bnj" event={"ID":"3727faad-272d-4a6a-a2ba-7b5be05cd11b","Type":"ContainerStarted","Data":"fe5fbf86df0368876aa815d6a76c6b9d94453af6aa6c53f5db8f896782593237"} Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.020724 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-52bnj" event={"ID":"3727faad-272d-4a6a-a2ba-7b5be05cd11b","Type":"ContainerStarted","Data":"f7d21193c0f145b336ef7657f21e89c667913e710a3dd574f4733fd59d881122"} Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.081963 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-create-52bnj" podStartSLOduration=2.081945012 podStartE2EDuration="2.081945012s" podCreationTimestamp="2025-12-09 10:14:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:14:24.053211054 +0000 UTC m=+1620.387715721" watchObservedRunningTime="2025-12-09 10:14:24.081945012 +0000 UTC m=+1620.416449679" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.224871 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-d6jn9"] Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.227418 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-d6jn9" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.233084 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.233117 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.261851 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-d6jn9"] Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.348356 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32-config-data\") pod \"nova-cell0-cell-mapping-d6jn9\" (UID: \"de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32\") " pod="openstack/nova-cell0-cell-mapping-d6jn9" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.348426 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32-scripts\") pod \"nova-cell0-cell-mapping-d6jn9\" (UID: \"de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32\") " pod="openstack/nova-cell0-cell-mapping-d6jn9" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.348482 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6t7m\" (UniqueName: \"kubernetes.io/projected/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32-kube-api-access-x6t7m\") pod \"nova-cell0-cell-mapping-d6jn9\" (UID: \"de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32\") " pod="openstack/nova-cell0-cell-mapping-d6jn9" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.348578 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-d6jn9\" (UID: \"de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32\") " pod="openstack/nova-cell0-cell-mapping-d6jn9" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.366851 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.369224 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.377426 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.393051 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.451213 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32-config-data\") pod \"nova-cell0-cell-mapping-d6jn9\" (UID: \"de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32\") " pod="openstack/nova-cell0-cell-mapping-d6jn9" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.451263 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32-scripts\") pod \"nova-cell0-cell-mapping-d6jn9\" (UID: \"de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32\") " pod="openstack/nova-cell0-cell-mapping-d6jn9" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.451308 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6t7m\" (UniqueName: \"kubernetes.io/projected/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32-kube-api-access-x6t7m\") pod \"nova-cell0-cell-mapping-d6jn9\" (UID: \"de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32\") " pod="openstack/nova-cell0-cell-mapping-d6jn9" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.451405 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-d6jn9\" (UID: \"de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32\") " pod="openstack/nova-cell0-cell-mapping-d6jn9" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.474644 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32-scripts\") pod \"nova-cell0-cell-mapping-d6jn9\" (UID: \"de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32\") " pod="openstack/nova-cell0-cell-mapping-d6jn9" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.476331 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-d6jn9\" (UID: \"de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32\") " pod="openstack/nova-cell0-cell-mapping-d6jn9" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.523123 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.528069 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6t7m\" (UniqueName: \"kubernetes.io/projected/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32-kube-api-access-x6t7m\") pod \"nova-cell0-cell-mapping-d6jn9\" (UID: \"de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32\") " pod="openstack/nova-cell0-cell-mapping-d6jn9" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.545403 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32-config-data\") pod \"nova-cell0-cell-mapping-d6jn9\" (UID: \"de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32\") " pod="openstack/nova-cell0-cell-mapping-d6jn9" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.546850 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.559374 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.559768 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhxd7\" (UniqueName: \"kubernetes.io/projected/0d38851d-cb31-4701-872e-ff5d8baa1c42-kube-api-access-nhxd7\") pod \"nova-scheduler-0\" (UID: \"0d38851d-cb31-4701-872e-ff5d8baa1c42\") " pod="openstack/nova-scheduler-0" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.563850 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d38851d-cb31-4701-872e-ff5d8baa1c42-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0d38851d-cb31-4701-872e-ff5d8baa1c42\") " pod="openstack/nova-scheduler-0" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.564019 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d38851d-cb31-4701-872e-ff5d8baa1c42-config-data\") pod \"nova-scheduler-0\" (UID: \"0d38851d-cb31-4701-872e-ff5d8baa1c42\") " pod="openstack/nova-scheduler-0" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.572995 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.583194 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-d6jn9" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.596506 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.614062 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.622828 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.626066 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 10:14:24 crc kubenswrapper[4824]: I1209 10:14:24.631275 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.126386 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d38851d-cb31-4701-872e-ff5d8baa1c42-config-data\") pod \"nova-scheduler-0\" (UID: \"0d38851d-cb31-4701-872e-ff5d8baa1c42\") " pod="openstack/nova-scheduler-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.126652 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9843016c-68f5-4f7a-b2a5-fe222d0ab28e-config-data\") pod \"nova-api-0\" (UID: \"9843016c-68f5-4f7a-b2a5-fe222d0ab28e\") " pod="openstack/nova-api-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.126706 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hf2j\" (UniqueName: \"kubernetes.io/projected/9843016c-68f5-4f7a-b2a5-fe222d0ab28e-kube-api-access-8hf2j\") pod \"nova-api-0\" (UID: \"9843016c-68f5-4f7a-b2a5-fe222d0ab28e\") " pod="openstack/nova-api-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.126818 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhxd7\" (UniqueName: \"kubernetes.io/projected/0d38851d-cb31-4701-872e-ff5d8baa1c42-kube-api-access-nhxd7\") pod \"nova-scheduler-0\" (UID: \"0d38851d-cb31-4701-872e-ff5d8baa1c42\") " pod="openstack/nova-scheduler-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.126904 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b97d1be8-3f68-488f-a879-ae2ab5b44cfb-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b97d1be8-3f68-488f-a879-ae2ab5b44cfb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.126939 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b97d1be8-3f68-488f-a879-ae2ab5b44cfb-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b97d1be8-3f68-488f-a879-ae2ab5b44cfb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.126959 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2kd5\" (UniqueName: \"kubernetes.io/projected/b97d1be8-3f68-488f-a879-ae2ab5b44cfb-kube-api-access-z2kd5\") pod \"nova-cell1-novncproxy-0\" (UID: \"b97d1be8-3f68-488f-a879-ae2ab5b44cfb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.127013 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9843016c-68f5-4f7a-b2a5-fe222d0ab28e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9843016c-68f5-4f7a-b2a5-fe222d0ab28e\") " pod="openstack/nova-api-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.127096 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d38851d-cb31-4701-872e-ff5d8baa1c42-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0d38851d-cb31-4701-872e-ff5d8baa1c42\") " pod="openstack/nova-scheduler-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.127125 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9843016c-68f5-4f7a-b2a5-fe222d0ab28e-logs\") pod \"nova-api-0\" (UID: \"9843016c-68f5-4f7a-b2a5-fe222d0ab28e\") " pod="openstack/nova-api-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.127869 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.131454 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.135732 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.153744 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d38851d-cb31-4701-872e-ff5d8baa1c42-config-data\") pod \"nova-scheduler-0\" (UID: \"0d38851d-cb31-4701-872e-ff5d8baa1c42\") " pod="openstack/nova-scheduler-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.164013 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d38851d-cb31-4701-872e-ff5d8baa1c42-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0d38851d-cb31-4701-872e-ff5d8baa1c42\") " pod="openstack/nova-scheduler-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.235986 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9843016c-68f5-4f7a-b2a5-fe222d0ab28e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9843016c-68f5-4f7a-b2a5-fe222d0ab28e\") " pod="openstack/nova-api-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.236112 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9843016c-68f5-4f7a-b2a5-fe222d0ab28e-logs\") pod \"nova-api-0\" (UID: \"9843016c-68f5-4f7a-b2a5-fe222d0ab28e\") " pod="openstack/nova-api-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.236379 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9843016c-68f5-4f7a-b2a5-fe222d0ab28e-config-data\") pod \"nova-api-0\" (UID: \"9843016c-68f5-4f7a-b2a5-fe222d0ab28e\") " pod="openstack/nova-api-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.236429 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hf2j\" (UniqueName: \"kubernetes.io/projected/9843016c-68f5-4f7a-b2a5-fe222d0ab28e-kube-api-access-8hf2j\") pod \"nova-api-0\" (UID: \"9843016c-68f5-4f7a-b2a5-fe222d0ab28e\") " pod="openstack/nova-api-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.236577 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b97d1be8-3f68-488f-a879-ae2ab5b44cfb-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b97d1be8-3f68-488f-a879-ae2ab5b44cfb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.236605 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b97d1be8-3f68-488f-a879-ae2ab5b44cfb-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b97d1be8-3f68-488f-a879-ae2ab5b44cfb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.236629 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2kd5\" (UniqueName: \"kubernetes.io/projected/b97d1be8-3f68-488f-a879-ae2ab5b44cfb-kube-api-access-z2kd5\") pod \"nova-cell1-novncproxy-0\" (UID: \"b97d1be8-3f68-488f-a879-ae2ab5b44cfb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.237910 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9843016c-68f5-4f7a-b2a5-fe222d0ab28e-logs\") pod \"nova-api-0\" (UID: \"9843016c-68f5-4f7a-b2a5-fe222d0ab28e\") " pod="openstack/nova-api-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.252685 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9843016c-68f5-4f7a-b2a5-fe222d0ab28e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9843016c-68f5-4f7a-b2a5-fe222d0ab28e\") " pod="openstack/nova-api-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.255241 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.258438 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9843016c-68f5-4f7a-b2a5-fe222d0ab28e-config-data\") pod \"nova-api-0\" (UID: \"9843016c-68f5-4f7a-b2a5-fe222d0ab28e\") " pod="openstack/nova-api-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.260497 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b97d1be8-3f68-488f-a879-ae2ab5b44cfb-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b97d1be8-3f68-488f-a879-ae2ab5b44cfb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.275755 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hf2j\" (UniqueName: \"kubernetes.io/projected/9843016c-68f5-4f7a-b2a5-fe222d0ab28e-kube-api-access-8hf2j\") pod \"nova-api-0\" (UID: \"9843016c-68f5-4f7a-b2a5-fe222d0ab28e\") " pod="openstack/nova-api-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.276373 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b97d1be8-3f68-488f-a879-ae2ab5b44cfb-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b97d1be8-3f68-488f-a879-ae2ab5b44cfb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.281237 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhxd7\" (UniqueName: \"kubernetes.io/projected/0d38851d-cb31-4701-872e-ff5d8baa1c42-kube-api-access-nhxd7\") pod \"nova-scheduler-0\" (UID: \"0d38851d-cb31-4701-872e-ff5d8baa1c42\") " pod="openstack/nova-scheduler-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.289929 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2kd5\" (UniqueName: \"kubernetes.io/projected/b97d1be8-3f68-488f-a879-ae2ab5b44cfb-kube-api-access-z2kd5\") pod \"nova-cell1-novncproxy-0\" (UID: \"b97d1be8-3f68-488f-a879-ae2ab5b44cfb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.297130 4824 generic.go:334] "Generic (PLEG): container finished" podID="3727faad-272d-4a6a-a2ba-7b5be05cd11b" containerID="fe5fbf86df0368876aa815d6a76c6b9d94453af6aa6c53f5db8f896782593237" exitCode=0 Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.303130 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-52bnj" event={"ID":"3727faad-272d-4a6a-a2ba-7b5be05cd11b","Type":"ContainerDied","Data":"fe5fbf86df0368876aa815d6a76c6b9d94453af6aa6c53f5db8f896782593237"} Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.464893 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-f925-account-create-update-zk59p" event={"ID":"96ede01d-246b-47e6-ac4d-06c74436006c","Type":"ContainerStarted","Data":"36892ea000779b2d91103ba0acc4ac4bc431aef8ae2a7a1f5876cf8a9e8eee23"} Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.480819 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk96j\" (UniqueName: \"kubernetes.io/projected/1707467c-0f4f-42a8-b964-910220950c75-kube-api-access-zk96j\") pod \"nova-metadata-0\" (UID: \"1707467c-0f4f-42a8-b964-910220950c75\") " pod="openstack/nova-metadata-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.480896 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1707467c-0f4f-42a8-b964-910220950c75-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1707467c-0f4f-42a8-b964-910220950c75\") " pod="openstack/nova-metadata-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.481096 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1707467c-0f4f-42a8-b964-910220950c75-logs\") pod \"nova-metadata-0\" (UID: \"1707467c-0f4f-42a8-b964-910220950c75\") " pod="openstack/nova-metadata-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.481508 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1707467c-0f4f-42a8-b964-910220950c75-config-data\") pod \"nova-metadata-0\" (UID: \"1707467c-0f4f-42a8-b964-910220950c75\") " pod="openstack/nova-metadata-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.501840 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.530862 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.542963 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc","Type":"ContainerStarted","Data":"fc355b2d4822ce3418c6d64d86a4f386510668e8f4fbf6826a4789f89c531b97"} Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.554918 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.579016 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-qcgm4"] Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.581288 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.584722 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk96j\" (UniqueName: \"kubernetes.io/projected/1707467c-0f4f-42a8-b964-910220950c75-kube-api-access-zk96j\") pod \"nova-metadata-0\" (UID: \"1707467c-0f4f-42a8-b964-910220950c75\") " pod="openstack/nova-metadata-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.584770 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1707467c-0f4f-42a8-b964-910220950c75-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1707467c-0f4f-42a8-b964-910220950c75\") " pod="openstack/nova-metadata-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.584856 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1707467c-0f4f-42a8-b964-910220950c75-logs\") pod \"nova-metadata-0\" (UID: \"1707467c-0f4f-42a8-b964-910220950c75\") " pod="openstack/nova-metadata-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.584952 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1707467c-0f4f-42a8-b964-910220950c75-config-data\") pod \"nova-metadata-0\" (UID: \"1707467c-0f4f-42a8-b964-910220950c75\") " pod="openstack/nova-metadata-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.585948 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1707467c-0f4f-42a8-b964-910220950c75-logs\") pod \"nova-metadata-0\" (UID: \"1707467c-0f4f-42a8-b964-910220950c75\") " pod="openstack/nova-metadata-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.612774 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1707467c-0f4f-42a8-b964-910220950c75-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1707467c-0f4f-42a8-b964-910220950c75\") " pod="openstack/nova-metadata-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.615805 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1707467c-0f4f-42a8-b964-910220950c75-config-data\") pod \"nova-metadata-0\" (UID: \"1707467c-0f4f-42a8-b964-910220950c75\") " pod="openstack/nova-metadata-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.624984 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-qcgm4"] Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.643085 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk96j\" (UniqueName: \"kubernetes.io/projected/1707467c-0f4f-42a8-b964-910220950c75-kube-api-access-zk96j\") pod \"nova-metadata-0\" (UID: \"1707467c-0f4f-42a8-b964-910220950c75\") " pod="openstack/nova-metadata-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.674313 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-f925-account-create-update-zk59p" podStartSLOduration=3.674289915 podStartE2EDuration="3.674289915s" podCreationTimestamp="2025-12-09 10:14:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:14:25.496851418 +0000 UTC m=+1621.831356085" watchObservedRunningTime="2025-12-09 10:14:25.674289915 +0000 UTC m=+1622.008794572" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.690119 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-qcgm4\" (UID: \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\") " pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.690230 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-config\") pod \"dnsmasq-dns-9b86998b5-qcgm4\" (UID: \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\") " pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.690349 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-qcgm4\" (UID: \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\") " pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.690391 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-qcgm4\" (UID: \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\") " pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.690435 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-dns-svc\") pod \"dnsmasq-dns-9b86998b5-qcgm4\" (UID: \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\") " pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.690474 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9nj5\" (UniqueName: \"kubernetes.io/projected/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-kube-api-access-x9nj5\") pod \"dnsmasq-dns-9b86998b5-qcgm4\" (UID: \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\") " pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.796601 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-qcgm4\" (UID: \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\") " pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.797240 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-qcgm4\" (UID: \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\") " pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.800537 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-qcgm4\" (UID: \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\") " pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.807336 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-dns-svc\") pod \"dnsmasq-dns-9b86998b5-qcgm4\" (UID: \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\") " pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.807620 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9nj5\" (UniqueName: \"kubernetes.io/projected/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-kube-api-access-x9nj5\") pod \"dnsmasq-dns-9b86998b5-qcgm4\" (UID: \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\") " pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.808212 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-qcgm4\" (UID: \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\") " pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.808292 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-config\") pod \"dnsmasq-dns-9b86998b5-qcgm4\" (UID: \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\") " pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.808697 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-qcgm4\" (UID: \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\") " pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.809516 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-dns-svc\") pod \"dnsmasq-dns-9b86998b5-qcgm4\" (UID: \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\") " pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.809877 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-config\") pod \"dnsmasq-dns-9b86998b5-qcgm4\" (UID: \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\") " pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.810136 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-qcgm4\" (UID: \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\") " pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.852634 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9nj5\" (UniqueName: \"kubernetes.io/projected/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-kube-api-access-x9nj5\") pod \"dnsmasq-dns-9b86998b5-qcgm4\" (UID: \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\") " pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.868252 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 10:14:25 crc kubenswrapper[4824]: I1209 10:14:25.912898 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" Dec 09 10:14:26 crc kubenswrapper[4824]: I1209 10:14:26.988213 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-d6jn9"] Dec 09 10:14:27 crc kubenswrapper[4824]: W1209 10:14:27.502004 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9843016c_68f5_4f7a_b2a5_fe222d0ab28e.slice/crio-2308b66baa8a7f643a046ca9bb78b35b295d7d05dad622b4c8ed4f873231548a WatchSource:0}: Error finding container 2308b66baa8a7f643a046ca9bb78b35b295d7d05dad622b4c8ed4f873231548a: Status 404 returned error can't find the container with id 2308b66baa8a7f643a046ca9bb78b35b295d7d05dad622b4c8ed4f873231548a Dec 09 10:14:27 crc kubenswrapper[4824]: I1209 10:14:27.525091 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 10:14:27 crc kubenswrapper[4824]: I1209 10:14:27.571869 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 10:14:27 crc kubenswrapper[4824]: I1209 10:14:27.606457 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 10:14:27 crc kubenswrapper[4824]: W1209 10:14:27.615506 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d38851d_cb31_4701_872e_ff5d8baa1c42.slice/crio-6a14b58f788c990f664be3a452a265ba4d25852931272770465624184400ac35 WatchSource:0}: Error finding container 6a14b58f788c990f664be3a452a265ba4d25852931272770465624184400ac35: Status 404 returned error can't find the container with id 6a14b58f788c990f664be3a452a265ba4d25852931272770465624184400ac35 Dec 09 10:14:27 crc kubenswrapper[4824]: I1209 10:14:27.633533 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 10:14:27 crc kubenswrapper[4824]: I1209 10:14:27.664863 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-d6jn9" event={"ID":"de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32","Type":"ContainerStarted","Data":"a263b2169fe43a5e644746c7440339c9006ef8de764e6dcbc5e1b876e301a6c7"} Dec 09 10:14:27 crc kubenswrapper[4824]: I1209 10:14:27.664925 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-d6jn9" event={"ID":"de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32","Type":"ContainerStarted","Data":"dd7f2f3f7c6c62441916fd2b9354605ee1bacf3c8f3c8b26dff5e21d52ac5c05"} Dec 09 10:14:27 crc kubenswrapper[4824]: I1209 10:14:27.697903 4824 generic.go:334] "Generic (PLEG): container finished" podID="96ede01d-246b-47e6-ac4d-06c74436006c" containerID="36892ea000779b2d91103ba0acc4ac4bc431aef8ae2a7a1f5876cf8a9e8eee23" exitCode=0 Dec 09 10:14:27 crc kubenswrapper[4824]: I1209 10:14:27.698073 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-f925-account-create-update-zk59p" event={"ID":"96ede01d-246b-47e6-ac4d-06c74436006c","Type":"ContainerDied","Data":"36892ea000779b2d91103ba0acc4ac4bc431aef8ae2a7a1f5876cf8a9e8eee23"} Dec 09 10:14:27 crc kubenswrapper[4824]: I1209 10:14:27.710144 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9843016c-68f5-4f7a-b2a5-fe222d0ab28e","Type":"ContainerStarted","Data":"2308b66baa8a7f643a046ca9bb78b35b295d7d05dad622b4c8ed4f873231548a"} Dec 09 10:14:27 crc kubenswrapper[4824]: I1209 10:14:27.728866 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-qcgm4"] Dec 09 10:14:27 crc kubenswrapper[4824]: I1209 10:14:27.731280 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc","Type":"ContainerStarted","Data":"7933eaa4a009a5ab39df47008d7965641e517ebececf33bf9eb8ecd85f7cdc0f"} Dec 09 10:14:27 crc kubenswrapper[4824]: I1209 10:14:27.760211 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1707467c-0f4f-42a8-b964-910220950c75","Type":"ContainerStarted","Data":"46599ac7926aef0dba29041551ef5f7adc203f8779386ca2a3b2859cbdb74d9b"} Dec 09 10:14:27 crc kubenswrapper[4824]: I1209 10:14:27.768582 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-d6jn9" podStartSLOduration=3.7685064759999998 podStartE2EDuration="3.768506476s" podCreationTimestamp="2025-12-09 10:14:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:14:27.718911889 +0000 UTC m=+1624.053416556" watchObservedRunningTime="2025-12-09 10:14:27.768506476 +0000 UTC m=+1624.103011143" Dec 09 10:14:27 crc kubenswrapper[4824]: I1209 10:14:27.800331 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-52bnj" Dec 09 10:14:27 crc kubenswrapper[4824]: I1209 10:14:27.906036 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3727faad-272d-4a6a-a2ba-7b5be05cd11b-operator-scripts\") pod \"3727faad-272d-4a6a-a2ba-7b5be05cd11b\" (UID: \"3727faad-272d-4a6a-a2ba-7b5be05cd11b\") " Dec 09 10:14:27 crc kubenswrapper[4824]: I1209 10:14:27.906178 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wn7n\" (UniqueName: \"kubernetes.io/projected/3727faad-272d-4a6a-a2ba-7b5be05cd11b-kube-api-access-4wn7n\") pod \"3727faad-272d-4a6a-a2ba-7b5be05cd11b\" (UID: \"3727faad-272d-4a6a-a2ba-7b5be05cd11b\") " Dec 09 10:14:27 crc kubenswrapper[4824]: I1209 10:14:27.907934 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3727faad-272d-4a6a-a2ba-7b5be05cd11b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3727faad-272d-4a6a-a2ba-7b5be05cd11b" (UID: "3727faad-272d-4a6a-a2ba-7b5be05cd11b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:14:27 crc kubenswrapper[4824]: I1209 10:14:27.914041 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3727faad-272d-4a6a-a2ba-7b5be05cd11b-kube-api-access-4wn7n" (OuterVolumeSpecName: "kube-api-access-4wn7n") pod "3727faad-272d-4a6a-a2ba-7b5be05cd11b" (UID: "3727faad-272d-4a6a-a2ba-7b5be05cd11b"). InnerVolumeSpecName "kube-api-access-4wn7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:14:28 crc kubenswrapper[4824]: I1209 10:14:28.016100 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3727faad-272d-4a6a-a2ba-7b5be05cd11b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:28 crc kubenswrapper[4824]: I1209 10:14:28.016937 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wn7n\" (UniqueName: \"kubernetes.io/projected/3727faad-272d-4a6a-a2ba-7b5be05cd11b-kube-api-access-4wn7n\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:29 crc kubenswrapper[4824]: I1209 10:14:29.009152 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc","Type":"ContainerStarted","Data":"6f481b7674fd6b82084202baf7a20c25a6ca710f3a119487376ed5729d517bd2"} Dec 09 10:14:29 crc kubenswrapper[4824]: I1209 10:14:29.020136 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0d38851d-cb31-4701-872e-ff5d8baa1c42","Type":"ContainerStarted","Data":"6a14b58f788c990f664be3a452a265ba4d25852931272770465624184400ac35"} Dec 09 10:14:29 crc kubenswrapper[4824]: I1209 10:14:29.027955 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-52bnj" Dec 09 10:14:29 crc kubenswrapper[4824]: I1209 10:14:29.028872 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-52bnj" event={"ID":"3727faad-272d-4a6a-a2ba-7b5be05cd11b","Type":"ContainerDied","Data":"f7d21193c0f145b336ef7657f21e89c667913e710a3dd574f4733fd59d881122"} Dec 09 10:14:29 crc kubenswrapper[4824]: I1209 10:14:29.028913 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7d21193c0f145b336ef7657f21e89c667913e710a3dd574f4733fd59d881122" Dec 09 10:14:29 crc kubenswrapper[4824]: I1209 10:14:29.040899 4824 generic.go:334] "Generic (PLEG): container finished" podID="b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6" containerID="1206eefde90c3313178d69b7b139b9d2fae3549a611ba3ca14e1378fa5e37c78" exitCode=0 Dec 09 10:14:29 crc kubenswrapper[4824]: I1209 10:14:29.041004 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" event={"ID":"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6","Type":"ContainerDied","Data":"1206eefde90c3313178d69b7b139b9d2fae3549a611ba3ca14e1378fa5e37c78"} Dec 09 10:14:29 crc kubenswrapper[4824]: I1209 10:14:29.041044 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" event={"ID":"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6","Type":"ContainerStarted","Data":"cdec60c4257e65e9647e5213309d6ec9812bacba891b84c2c517eff535223385"} Dec 09 10:14:29 crc kubenswrapper[4824]: I1209 10:14:29.048478 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b97d1be8-3f68-488f-a879-ae2ab5b44cfb","Type":"ContainerStarted","Data":"ff86603f59e1e5c9da9104481b88a3201327044676b4755fc48536340584549a"} Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.054325 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rcv9n"] Dec 09 10:14:30 crc kubenswrapper[4824]: E1209 10:14:30.055376 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3727faad-272d-4a6a-a2ba-7b5be05cd11b" containerName="mariadb-database-create" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.055426 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3727faad-272d-4a6a-a2ba-7b5be05cd11b" containerName="mariadb-database-create" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.055759 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3727faad-272d-4a6a-a2ba-7b5be05cd11b" containerName="mariadb-database-create" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.056832 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rcv9n" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.062182 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.062496 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.085509 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rcv9n"] Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.119247 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" event={"ID":"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6","Type":"ContainerStarted","Data":"b699134baa3469d213951ac530d09148ea90fdf79e2fe5574dc7eb9184ca9b37"} Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.120427 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.153310 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" podStartSLOduration=5.153292107 podStartE2EDuration="5.153292107s" podCreationTimestamp="2025-12-09 10:14:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:14:30.145908584 +0000 UTC m=+1626.480413261" watchObservedRunningTime="2025-12-09 10:14:30.153292107 +0000 UTC m=+1626.487796764" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.173729 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5f03343-b401-422a-bea2-7c97f7a99600-config-data\") pod \"nova-cell1-conductor-db-sync-rcv9n\" (UID: \"b5f03343-b401-422a-bea2-7c97f7a99600\") " pod="openstack/nova-cell1-conductor-db-sync-rcv9n" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.173908 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7x4d\" (UniqueName: \"kubernetes.io/projected/b5f03343-b401-422a-bea2-7c97f7a99600-kube-api-access-c7x4d\") pod \"nova-cell1-conductor-db-sync-rcv9n\" (UID: \"b5f03343-b401-422a-bea2-7c97f7a99600\") " pod="openstack/nova-cell1-conductor-db-sync-rcv9n" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.174027 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5f03343-b401-422a-bea2-7c97f7a99600-scripts\") pod \"nova-cell1-conductor-db-sync-rcv9n\" (UID: \"b5f03343-b401-422a-bea2-7c97f7a99600\") " pod="openstack/nova-cell1-conductor-db-sync-rcv9n" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.174206 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5f03343-b401-422a-bea2-7c97f7a99600-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-rcv9n\" (UID: \"b5f03343-b401-422a-bea2-7c97f7a99600\") " pod="openstack/nova-cell1-conductor-db-sync-rcv9n" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.277204 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7x4d\" (UniqueName: \"kubernetes.io/projected/b5f03343-b401-422a-bea2-7c97f7a99600-kube-api-access-c7x4d\") pod \"nova-cell1-conductor-db-sync-rcv9n\" (UID: \"b5f03343-b401-422a-bea2-7c97f7a99600\") " pod="openstack/nova-cell1-conductor-db-sync-rcv9n" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.277523 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5f03343-b401-422a-bea2-7c97f7a99600-scripts\") pod \"nova-cell1-conductor-db-sync-rcv9n\" (UID: \"b5f03343-b401-422a-bea2-7c97f7a99600\") " pod="openstack/nova-cell1-conductor-db-sync-rcv9n" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.277689 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5f03343-b401-422a-bea2-7c97f7a99600-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-rcv9n\" (UID: \"b5f03343-b401-422a-bea2-7c97f7a99600\") " pod="openstack/nova-cell1-conductor-db-sync-rcv9n" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.277736 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5f03343-b401-422a-bea2-7c97f7a99600-config-data\") pod \"nova-cell1-conductor-db-sync-rcv9n\" (UID: \"b5f03343-b401-422a-bea2-7c97f7a99600\") " pod="openstack/nova-cell1-conductor-db-sync-rcv9n" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.288404 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5f03343-b401-422a-bea2-7c97f7a99600-config-data\") pod \"nova-cell1-conductor-db-sync-rcv9n\" (UID: \"b5f03343-b401-422a-bea2-7c97f7a99600\") " pod="openstack/nova-cell1-conductor-db-sync-rcv9n" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.327553 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7x4d\" (UniqueName: \"kubernetes.io/projected/b5f03343-b401-422a-bea2-7c97f7a99600-kube-api-access-c7x4d\") pod \"nova-cell1-conductor-db-sync-rcv9n\" (UID: \"b5f03343-b401-422a-bea2-7c97f7a99600\") " pod="openstack/nova-cell1-conductor-db-sync-rcv9n" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.328636 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5f03343-b401-422a-bea2-7c97f7a99600-scripts\") pod \"nova-cell1-conductor-db-sync-rcv9n\" (UID: \"b5f03343-b401-422a-bea2-7c97f7a99600\") " pod="openstack/nova-cell1-conductor-db-sync-rcv9n" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.328984 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5f03343-b401-422a-bea2-7c97f7a99600-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-rcv9n\" (UID: \"b5f03343-b401-422a-bea2-7c97f7a99600\") " pod="openstack/nova-cell1-conductor-db-sync-rcv9n" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.472045 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rcv9n" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.477037 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-f925-account-create-update-zk59p" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.585810 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgkx6\" (UniqueName: \"kubernetes.io/projected/96ede01d-246b-47e6-ac4d-06c74436006c-kube-api-access-fgkx6\") pod \"96ede01d-246b-47e6-ac4d-06c74436006c\" (UID: \"96ede01d-246b-47e6-ac4d-06c74436006c\") " Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.585926 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96ede01d-246b-47e6-ac4d-06c74436006c-operator-scripts\") pod \"96ede01d-246b-47e6-ac4d-06c74436006c\" (UID: \"96ede01d-246b-47e6-ac4d-06c74436006c\") " Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.587232 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96ede01d-246b-47e6-ac4d-06c74436006c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "96ede01d-246b-47e6-ac4d-06c74436006c" (UID: "96ede01d-246b-47e6-ac4d-06c74436006c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.604519 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96ede01d-246b-47e6-ac4d-06c74436006c-kube-api-access-fgkx6" (OuterVolumeSpecName: "kube-api-access-fgkx6") pod "96ede01d-246b-47e6-ac4d-06c74436006c" (UID: "96ede01d-246b-47e6-ac4d-06c74436006c"). InnerVolumeSpecName "kube-api-access-fgkx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.691879 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgkx6\" (UniqueName: \"kubernetes.io/projected/96ede01d-246b-47e6-ac4d-06c74436006c-kube-api-access-fgkx6\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:30 crc kubenswrapper[4824]: I1209 10:14:30.691915 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96ede01d-246b-47e6-ac4d-06c74436006c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:31 crc kubenswrapper[4824]: I1209 10:14:31.007462 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 10:14:31 crc kubenswrapper[4824]: I1209 10:14:31.025518 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 10:14:31 crc kubenswrapper[4824]: I1209 10:14:31.158882 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-f925-account-create-update-zk59p" Dec 09 10:14:31 crc kubenswrapper[4824]: I1209 10:14:31.158972 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-f925-account-create-update-zk59p" event={"ID":"96ede01d-246b-47e6-ac4d-06c74436006c","Type":"ContainerDied","Data":"c7e00fbabf965185f45930f286488d6096d87daf00d7fff9406b47db876962c8"} Dec 09 10:14:31 crc kubenswrapper[4824]: I1209 10:14:31.159032 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7e00fbabf965185f45930f286488d6096d87daf00d7fff9406b47db876962c8" Dec 09 10:14:31 crc kubenswrapper[4824]: I1209 10:14:31.205962 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc","Type":"ContainerStarted","Data":"f0052a0f9a94b77f605971d2f4d65e66b5cb33beee7c8b898c44b20df19c263c"} Dec 09 10:14:31 crc kubenswrapper[4824]: I1209 10:14:31.255891 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rcv9n"] Dec 09 10:14:32 crc kubenswrapper[4824]: I1209 10:14:32.224427 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rcv9n" event={"ID":"b5f03343-b401-422a-bea2-7c97f7a99600","Type":"ContainerStarted","Data":"4963e37b81088e0503441fed1cac51f16773707d7ec7f43b13074e3262d3cd1b"} Dec 09 10:14:32 crc kubenswrapper[4824]: I1209 10:14:32.721275 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-jvpcc"] Dec 09 10:14:32 crc kubenswrapper[4824]: E1209 10:14:32.722102 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96ede01d-246b-47e6-ac4d-06c74436006c" containerName="mariadb-account-create-update" Dec 09 10:14:32 crc kubenswrapper[4824]: I1209 10:14:32.722117 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="96ede01d-246b-47e6-ac4d-06c74436006c" containerName="mariadb-account-create-update" Dec 09 10:14:32 crc kubenswrapper[4824]: I1209 10:14:32.722345 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="96ede01d-246b-47e6-ac4d-06c74436006c" containerName="mariadb-account-create-update" Dec 09 10:14:32 crc kubenswrapper[4824]: I1209 10:14:32.723478 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-jvpcc" Dec 09 10:14:32 crc kubenswrapper[4824]: I1209 10:14:32.729564 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 09 10:14:32 crc kubenswrapper[4824]: I1209 10:14:32.731177 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-mgzrr" Dec 09 10:14:32 crc kubenswrapper[4824]: I1209 10:14:32.731520 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 09 10:14:32 crc kubenswrapper[4824]: I1209 10:14:32.731889 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 09 10:14:32 crc kubenswrapper[4824]: I1209 10:14:32.737452 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-jvpcc"] Dec 09 10:14:32 crc kubenswrapper[4824]: I1209 10:14:32.828959 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4eb5fcc0-0113-4709-881b-7d109571a08d-scripts\") pod \"aodh-db-sync-jvpcc\" (UID: \"4eb5fcc0-0113-4709-881b-7d109571a08d\") " pod="openstack/aodh-db-sync-jvpcc" Dec 09 10:14:32 crc kubenswrapper[4824]: I1209 10:14:32.829064 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eb5fcc0-0113-4709-881b-7d109571a08d-config-data\") pod \"aodh-db-sync-jvpcc\" (UID: \"4eb5fcc0-0113-4709-881b-7d109571a08d\") " pod="openstack/aodh-db-sync-jvpcc" Dec 09 10:14:32 crc kubenswrapper[4824]: I1209 10:14:32.829117 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wdph\" (UniqueName: \"kubernetes.io/projected/4eb5fcc0-0113-4709-881b-7d109571a08d-kube-api-access-6wdph\") pod \"aodh-db-sync-jvpcc\" (UID: \"4eb5fcc0-0113-4709-881b-7d109571a08d\") " pod="openstack/aodh-db-sync-jvpcc" Dec 09 10:14:32 crc kubenswrapper[4824]: I1209 10:14:32.829145 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eb5fcc0-0113-4709-881b-7d109571a08d-combined-ca-bundle\") pod \"aodh-db-sync-jvpcc\" (UID: \"4eb5fcc0-0113-4709-881b-7d109571a08d\") " pod="openstack/aodh-db-sync-jvpcc" Dec 09 10:14:32 crc kubenswrapper[4824]: I1209 10:14:32.860589 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:14:32 crc kubenswrapper[4824]: I1209 10:14:32.860692 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:14:32 crc kubenswrapper[4824]: I1209 10:14:32.933110 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4eb5fcc0-0113-4709-881b-7d109571a08d-scripts\") pod \"aodh-db-sync-jvpcc\" (UID: \"4eb5fcc0-0113-4709-881b-7d109571a08d\") " pod="openstack/aodh-db-sync-jvpcc" Dec 09 10:14:32 crc kubenswrapper[4824]: I1209 10:14:32.933248 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eb5fcc0-0113-4709-881b-7d109571a08d-config-data\") pod \"aodh-db-sync-jvpcc\" (UID: \"4eb5fcc0-0113-4709-881b-7d109571a08d\") " pod="openstack/aodh-db-sync-jvpcc" Dec 09 10:14:32 crc kubenswrapper[4824]: I1209 10:14:32.933309 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wdph\" (UniqueName: \"kubernetes.io/projected/4eb5fcc0-0113-4709-881b-7d109571a08d-kube-api-access-6wdph\") pod \"aodh-db-sync-jvpcc\" (UID: \"4eb5fcc0-0113-4709-881b-7d109571a08d\") " pod="openstack/aodh-db-sync-jvpcc" Dec 09 10:14:32 crc kubenswrapper[4824]: I1209 10:14:32.933335 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eb5fcc0-0113-4709-881b-7d109571a08d-combined-ca-bundle\") pod \"aodh-db-sync-jvpcc\" (UID: \"4eb5fcc0-0113-4709-881b-7d109571a08d\") " pod="openstack/aodh-db-sync-jvpcc" Dec 09 10:14:32 crc kubenswrapper[4824]: I1209 10:14:32.940979 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4eb5fcc0-0113-4709-881b-7d109571a08d-scripts\") pod \"aodh-db-sync-jvpcc\" (UID: \"4eb5fcc0-0113-4709-881b-7d109571a08d\") " pod="openstack/aodh-db-sync-jvpcc" Dec 09 10:14:32 crc kubenswrapper[4824]: I1209 10:14:32.950021 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eb5fcc0-0113-4709-881b-7d109571a08d-config-data\") pod \"aodh-db-sync-jvpcc\" (UID: \"4eb5fcc0-0113-4709-881b-7d109571a08d\") " pod="openstack/aodh-db-sync-jvpcc" Dec 09 10:14:32 crc kubenswrapper[4824]: I1209 10:14:32.968376 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eb5fcc0-0113-4709-881b-7d109571a08d-combined-ca-bundle\") pod \"aodh-db-sync-jvpcc\" (UID: \"4eb5fcc0-0113-4709-881b-7d109571a08d\") " pod="openstack/aodh-db-sync-jvpcc" Dec 09 10:14:33 crc kubenswrapper[4824]: I1209 10:14:33.005353 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wdph\" (UniqueName: \"kubernetes.io/projected/4eb5fcc0-0113-4709-881b-7d109571a08d-kube-api-access-6wdph\") pod \"aodh-db-sync-jvpcc\" (UID: \"4eb5fcc0-0113-4709-881b-7d109571a08d\") " pod="openstack/aodh-db-sync-jvpcc" Dec 09 10:14:33 crc kubenswrapper[4824]: I1209 10:14:33.102112 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-jvpcc" Dec 09 10:14:33 crc kubenswrapper[4824]: I1209 10:14:33.242016 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rcv9n" event={"ID":"b5f03343-b401-422a-bea2-7c97f7a99600","Type":"ContainerStarted","Data":"2c46216c2b5f4579d9654864a4f17ec06cebbdf5bcccad4f0d54c87f26e021c3"} Dec 09 10:14:33 crc kubenswrapper[4824]: I1209 10:14:33.279285 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-rcv9n" podStartSLOduration=3.279260628 podStartE2EDuration="3.279260628s" podCreationTimestamp="2025-12-09 10:14:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:14:33.264364598 +0000 UTC m=+1629.598869265" watchObservedRunningTime="2025-12-09 10:14:33.279260628 +0000 UTC m=+1629.613765285" Dec 09 10:14:34 crc kubenswrapper[4824]: I1209 10:14:34.260347 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc","Type":"ContainerStarted","Data":"0a9b3999d41865e7b6906497b24f006acab61039d97dccd40fe6172382468d54"} Dec 09 10:14:34 crc kubenswrapper[4824]: I1209 10:14:34.287114 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.887414465 podStartE2EDuration="12.287093862s" podCreationTimestamp="2025-12-09 10:14:22 +0000 UTC" firstStartedPulling="2025-12-09 10:14:24.571229372 +0000 UTC m=+1620.905734039" lastFinishedPulling="2025-12-09 10:14:31.970908769 +0000 UTC m=+1628.305413436" observedRunningTime="2025-12-09 10:14:34.286413631 +0000 UTC m=+1630.620918298" watchObservedRunningTime="2025-12-09 10:14:34.287093862 +0000 UTC m=+1630.621598529" Dec 09 10:14:35 crc kubenswrapper[4824]: I1209 10:14:35.270914 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 10:14:35 crc kubenswrapper[4824]: I1209 10:14:35.927420 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" Dec 09 10:14:36 crc kubenswrapper[4824]: I1209 10:14:36.040371 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-tcz7b"] Dec 09 10:14:36 crc kubenswrapper[4824]: I1209 10:14:36.040873 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" podUID="468592b2-a40d-47d8-aaf5-0b77b3f4e0a8" containerName="dnsmasq-dns" containerID="cri-o://30c1dfbbbf33f5341b9797eb0f33e46c09adec7ea6e20d991ab0e82e9986f5a7" gracePeriod=10 Dec 09 10:14:36 crc kubenswrapper[4824]: I1209 10:14:36.301363 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" podUID="468592b2-a40d-47d8-aaf5-0b77b3f4e0a8" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.211:5353: connect: connection refused" Dec 09 10:14:36 crc kubenswrapper[4824]: I1209 10:14:36.339609 4824 generic.go:334] "Generic (PLEG): container finished" podID="468592b2-a40d-47d8-aaf5-0b77b3f4e0a8" containerID="30c1dfbbbf33f5341b9797eb0f33e46c09adec7ea6e20d991ab0e82e9986f5a7" exitCode=0 Dec 09 10:14:36 crc kubenswrapper[4824]: I1209 10:14:36.341385 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" event={"ID":"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8","Type":"ContainerDied","Data":"30c1dfbbbf33f5341b9797eb0f33e46c09adec7ea6e20d991ab0e82e9986f5a7"} Dec 09 10:14:37 crc kubenswrapper[4824]: I1209 10:14:37.925746 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.139357 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-config\") pod \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\" (UID: \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\") " Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.139441 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l79k7\" (UniqueName: \"kubernetes.io/projected/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-kube-api-access-l79k7\") pod \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\" (UID: \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\") " Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.139510 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-dns-swift-storage-0\") pod \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\" (UID: \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\") " Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.139752 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-dns-svc\") pod \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\" (UID: \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\") " Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.139841 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-ovsdbserver-sb\") pod \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\" (UID: \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\") " Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.139912 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-ovsdbserver-nb\") pod \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\" (UID: \"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8\") " Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.219862 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-jvpcc"] Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.307379 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-kube-api-access-l79k7" (OuterVolumeSpecName: "kube-api-access-l79k7") pod "468592b2-a40d-47d8-aaf5-0b77b3f4e0a8" (UID: "468592b2-a40d-47d8-aaf5-0b77b3f4e0a8"). InnerVolumeSpecName "kube-api-access-l79k7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.318393 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "468592b2-a40d-47d8-aaf5-0b77b3f4e0a8" (UID: "468592b2-a40d-47d8-aaf5-0b77b3f4e0a8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.361613 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.361644 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l79k7\" (UniqueName: \"kubernetes.io/projected/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-kube-api-access-l79k7\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.377987 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "468592b2-a40d-47d8-aaf5-0b77b3f4e0a8" (UID: "468592b2-a40d-47d8-aaf5-0b77b3f4e0a8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.399163 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b97d1be8-3f68-488f-a879-ae2ab5b44cfb","Type":"ContainerStarted","Data":"671b4caf32e6e41880e92594a0ad626b69af708fa7f1ccd387e7fdd7f87aaa0f"} Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.399341 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="b97d1be8-3f68-488f-a879-ae2ab5b44cfb" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://671b4caf32e6e41880e92594a0ad626b69af708fa7f1ccd387e7fdd7f87aaa0f" gracePeriod=30 Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.414029 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-jvpcc" event={"ID":"4eb5fcc0-0113-4709-881b-7d109571a08d","Type":"ContainerStarted","Data":"f313b8129a47b7e4efe43b22e6737ee535fbba1bcc54ef995d3ac7edd41d93e4"} Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.422637 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=5.082481613 podStartE2EDuration="14.422616472s" podCreationTimestamp="2025-12-09 10:14:24 +0000 UTC" firstStartedPulling="2025-12-09 10:14:27.649195455 +0000 UTC m=+1623.983700122" lastFinishedPulling="2025-12-09 10:14:36.989330314 +0000 UTC m=+1633.323834981" observedRunningTime="2025-12-09 10:14:38.421740174 +0000 UTC m=+1634.756244841" watchObservedRunningTime="2025-12-09 10:14:38.422616472 +0000 UTC m=+1634.757121139" Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.422987 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "468592b2-a40d-47d8-aaf5-0b77b3f4e0a8" (UID: "468592b2-a40d-47d8-aaf5-0b77b3f4e0a8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.424547 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0d38851d-cb31-4701-872e-ff5d8baa1c42","Type":"ContainerStarted","Data":"0c0c7ca9f25255d4eb200c86bc0d5898d5bda84a3dd50cd6091a12920be7f548"} Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.430325 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1707467c-0f4f-42a8-b964-910220950c75","Type":"ContainerStarted","Data":"7fb5ed24eef96c26707affd19903c01d8e63b2ecd4500ada281bf040b581a3e6"} Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.438438 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.438473 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-tcz7b" event={"ID":"468592b2-a40d-47d8-aaf5-0b77b3f4e0a8","Type":"ContainerDied","Data":"c5c363f24710905cd95df44fce88db788c882b7a75ecece1756df8edb8e99e01"} Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.438538 4824 scope.go:117] "RemoveContainer" containerID="30c1dfbbbf33f5341b9797eb0f33e46c09adec7ea6e20d991ab0e82e9986f5a7" Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.468882 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.468930 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.472516 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-config" (OuterVolumeSpecName: "config") pod "468592b2-a40d-47d8-aaf5-0b77b3f4e0a8" (UID: "468592b2-a40d-47d8-aaf5-0b77b3f4e0a8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.473834 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "468592b2-a40d-47d8-aaf5-0b77b3f4e0a8" (UID: "468592b2-a40d-47d8-aaf5-0b77b3f4e0a8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.488887 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9843016c-68f5-4f7a-b2a5-fe222d0ab28e","Type":"ContainerStarted","Data":"07ff2b0704a6589199b5b0d22f647e6a6a4a5fa97bbc72b662fd861ededc4fad"} Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.511084 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=5.14912444 podStartE2EDuration="14.511058267s" podCreationTimestamp="2025-12-09 10:14:24 +0000 UTC" firstStartedPulling="2025-12-09 10:14:27.656125405 +0000 UTC m=+1623.990630072" lastFinishedPulling="2025-12-09 10:14:37.018059232 +0000 UTC m=+1633.352563899" observedRunningTime="2025-12-09 10:14:38.474403878 +0000 UTC m=+1634.808908545" watchObservedRunningTime="2025-12-09 10:14:38.511058267 +0000 UTC m=+1634.845562934" Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.549752 4824 scope.go:117] "RemoveContainer" containerID="ccfba621e765d76b3511a4bae1bf6a4529bcb8b04d6255c797510fca2c5e97b8" Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.570983 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.571015 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.818849 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-tcz7b"] Dec 09 10:14:38 crc kubenswrapper[4824]: I1209 10:14:38.831094 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-tcz7b"] Dec 09 10:14:40 crc kubenswrapper[4824]: I1209 10:14:40.133587 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="468592b2-a40d-47d8-aaf5-0b77b3f4e0a8" path="/var/lib/kubelet/pods/468592b2-a40d-47d8-aaf5-0b77b3f4e0a8/volumes" Dec 09 10:14:40 crc kubenswrapper[4824]: I1209 10:14:40.163115 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1707467c-0f4f-42a8-b964-910220950c75","Type":"ContainerStarted","Data":"c8b9a4233646a244b18f60c62c548a55eb0d5f9cd304d4c2db994370c2ad4108"} Dec 09 10:14:40 crc kubenswrapper[4824]: I1209 10:14:40.163348 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1707467c-0f4f-42a8-b964-910220950c75" containerName="nova-metadata-log" containerID="cri-o://7fb5ed24eef96c26707affd19903c01d8e63b2ecd4500ada281bf040b581a3e6" gracePeriod=30 Dec 09 10:14:40 crc kubenswrapper[4824]: I1209 10:14:40.163446 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1707467c-0f4f-42a8-b964-910220950c75" containerName="nova-metadata-metadata" containerID="cri-o://c8b9a4233646a244b18f60c62c548a55eb0d5f9cd304d4c2db994370c2ad4108" gracePeriod=30 Dec 09 10:14:40 crc kubenswrapper[4824]: I1209 10:14:40.262808 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=6.70817583 podStartE2EDuration="16.262766775s" podCreationTimestamp="2025-12-09 10:14:24 +0000 UTC" firstStartedPulling="2025-12-09 10:14:27.537719863 +0000 UTC m=+1623.872224540" lastFinishedPulling="2025-12-09 10:14:37.092310828 +0000 UTC m=+1633.426815485" observedRunningTime="2025-12-09 10:14:40.25470895 +0000 UTC m=+1636.589213617" watchObservedRunningTime="2025-12-09 10:14:40.262766775 +0000 UTC m=+1636.597271442" Dec 09 10:14:40 crc kubenswrapper[4824]: I1209 10:14:40.266498 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9843016c-68f5-4f7a-b2a5-fe222d0ab28e","Type":"ContainerStarted","Data":"2fab16a089078981ace99988cb78583d5c3f2eaa3a9b19c76e2869be5cf6ee15"} Dec 09 10:14:40 crc kubenswrapper[4824]: I1209 10:14:40.367614 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=6.906778604 podStartE2EDuration="16.367581666s" podCreationTimestamp="2025-12-09 10:14:24 +0000 UTC" firstStartedPulling="2025-12-09 10:14:27.513141027 +0000 UTC m=+1623.847645694" lastFinishedPulling="2025-12-09 10:14:36.973944089 +0000 UTC m=+1633.308448756" observedRunningTime="2025-12-09 10:14:40.364486149 +0000 UTC m=+1636.698990826" watchObservedRunningTime="2025-12-09 10:14:40.367581666 +0000 UTC m=+1636.702086333" Dec 09 10:14:40 crc kubenswrapper[4824]: I1209 10:14:40.506677 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 09 10:14:40 crc kubenswrapper[4824]: I1209 10:14:40.532928 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:14:40 crc kubenswrapper[4824]: I1209 10:14:40.938233 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 09 10:14:40 crc kubenswrapper[4824]: I1209 10:14:40.938322 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 09 10:14:40 crc kubenswrapper[4824]: I1209 10:14:40.940719 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="73d6bd70-44c7-4eed-a93a-36df636869cf" containerName="galera" probeResult="failure" output="command timed out" Dec 09 10:14:40 crc kubenswrapper[4824]: I1209 10:14:40.987367 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="73d6bd70-44c7-4eed-a93a-36df636869cf" containerName="galera" probeResult="failure" output="command timed out" Dec 09 10:14:41 crc kubenswrapper[4824]: I1209 10:14:41.298406 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1707467c-0f4f-42a8-b964-910220950c75","Type":"ContainerDied","Data":"7fb5ed24eef96c26707affd19903c01d8e63b2ecd4500ada281bf040b581a3e6"} Dec 09 10:14:41 crc kubenswrapper[4824]: I1209 10:14:41.298793 4824 generic.go:334] "Generic (PLEG): container finished" podID="1707467c-0f4f-42a8-b964-910220950c75" containerID="7fb5ed24eef96c26707affd19903c01d8e63b2ecd4500ada281bf040b581a3e6" exitCode=143 Dec 09 10:14:42 crc kubenswrapper[4824]: I1209 10:14:42.356335 4824 generic.go:334] "Generic (PLEG): container finished" podID="1707467c-0f4f-42a8-b964-910220950c75" containerID="c8b9a4233646a244b18f60c62c548a55eb0d5f9cd304d4c2db994370c2ad4108" exitCode=0 Dec 09 10:14:42 crc kubenswrapper[4824]: I1209 10:14:42.356422 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1707467c-0f4f-42a8-b964-910220950c75","Type":"ContainerDied","Data":"c8b9a4233646a244b18f60c62c548a55eb0d5f9cd304d4c2db994370c2ad4108"} Dec 09 10:14:42 crc kubenswrapper[4824]: I1209 10:14:42.923137 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.215288 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zk96j\" (UniqueName: \"kubernetes.io/projected/1707467c-0f4f-42a8-b964-910220950c75-kube-api-access-zk96j\") pod \"1707467c-0f4f-42a8-b964-910220950c75\" (UID: \"1707467c-0f4f-42a8-b964-910220950c75\") " Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.215470 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1707467c-0f4f-42a8-b964-910220950c75-combined-ca-bundle\") pod \"1707467c-0f4f-42a8-b964-910220950c75\" (UID: \"1707467c-0f4f-42a8-b964-910220950c75\") " Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.215592 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1707467c-0f4f-42a8-b964-910220950c75-config-data\") pod \"1707467c-0f4f-42a8-b964-910220950c75\" (UID: \"1707467c-0f4f-42a8-b964-910220950c75\") " Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.215653 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1707467c-0f4f-42a8-b964-910220950c75-logs\") pod \"1707467c-0f4f-42a8-b964-910220950c75\" (UID: \"1707467c-0f4f-42a8-b964-910220950c75\") " Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.217030 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1707467c-0f4f-42a8-b964-910220950c75-logs" (OuterVolumeSpecName: "logs") pod "1707467c-0f4f-42a8-b964-910220950c75" (UID: "1707467c-0f4f-42a8-b964-910220950c75"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.265139 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1707467c-0f4f-42a8-b964-910220950c75-kube-api-access-zk96j" (OuterVolumeSpecName: "kube-api-access-zk96j") pod "1707467c-0f4f-42a8-b964-910220950c75" (UID: "1707467c-0f4f-42a8-b964-910220950c75"). InnerVolumeSpecName "kube-api-access-zk96j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.295159 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1707467c-0f4f-42a8-b964-910220950c75-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1707467c-0f4f-42a8-b964-910220950c75" (UID: "1707467c-0f4f-42a8-b964-910220950c75"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.306226 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1707467c-0f4f-42a8-b964-910220950c75-config-data" (OuterVolumeSpecName: "config-data") pod "1707467c-0f4f-42a8-b964-910220950c75" (UID: "1707467c-0f4f-42a8-b964-910220950c75"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.327646 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1707467c-0f4f-42a8-b964-910220950c75-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.327707 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1707467c-0f4f-42a8-b964-910220950c75-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.327723 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1707467c-0f4f-42a8-b964-910220950c75-logs\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.327737 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zk96j\" (UniqueName: \"kubernetes.io/projected/1707467c-0f4f-42a8-b964-910220950c75-kube-api-access-zk96j\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.392976 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1707467c-0f4f-42a8-b964-910220950c75","Type":"ContainerDied","Data":"46599ac7926aef0dba29041551ef5f7adc203f8779386ca2a3b2859cbdb74d9b"} Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.393078 4824 scope.go:117] "RemoveContainer" containerID="c8b9a4233646a244b18f60c62c548a55eb0d5f9cd304d4c2db994370c2ad4108" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.393400 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.472921 4824 scope.go:117] "RemoveContainer" containerID="7fb5ed24eef96c26707affd19903c01d8e63b2ecd4500ada281bf040b581a3e6" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.481003 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.513849 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.551977 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 09 10:14:43 crc kubenswrapper[4824]: E1209 10:14:43.552674 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1707467c-0f4f-42a8-b964-910220950c75" containerName="nova-metadata-metadata" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.552692 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1707467c-0f4f-42a8-b964-910220950c75" containerName="nova-metadata-metadata" Dec 09 10:14:43 crc kubenswrapper[4824]: E1209 10:14:43.552712 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="468592b2-a40d-47d8-aaf5-0b77b3f4e0a8" containerName="dnsmasq-dns" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.552719 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="468592b2-a40d-47d8-aaf5-0b77b3f4e0a8" containerName="dnsmasq-dns" Dec 09 10:14:43 crc kubenswrapper[4824]: E1209 10:14:43.552768 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="468592b2-a40d-47d8-aaf5-0b77b3f4e0a8" containerName="init" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.552800 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="468592b2-a40d-47d8-aaf5-0b77b3f4e0a8" containerName="init" Dec 09 10:14:43 crc kubenswrapper[4824]: E1209 10:14:43.552825 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1707467c-0f4f-42a8-b964-910220950c75" containerName="nova-metadata-log" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.552832 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1707467c-0f4f-42a8-b964-910220950c75" containerName="nova-metadata-log" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.553088 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1707467c-0f4f-42a8-b964-910220950c75" containerName="nova-metadata-metadata" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.553106 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1707467c-0f4f-42a8-b964-910220950c75" containerName="nova-metadata-log" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.553117 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="468592b2-a40d-47d8-aaf5-0b77b3f4e0a8" containerName="dnsmasq-dns" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.554530 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.573264 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.573901 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.579876 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.833231 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\") " pod="openstack/nova-metadata-0" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.833369 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-logs\") pod \"nova-metadata-0\" (UID: \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\") " pod="openstack/nova-metadata-0" Dec 09 10:14:43 crc kubenswrapper[4824]: I1209 10:14:43.833435 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ck8j9\" (UniqueName: \"kubernetes.io/projected/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-kube-api-access-ck8j9\") pod \"nova-metadata-0\" (UID: \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\") " pod="openstack/nova-metadata-0" Dec 09 10:14:44 crc kubenswrapper[4824]: I1209 10:14:44.001080 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-config-data\") pod \"nova-metadata-0\" (UID: \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\") " pod="openstack/nova-metadata-0" Dec 09 10:14:44 crc kubenswrapper[4824]: I1209 10:14:44.001198 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\") " pod="openstack/nova-metadata-0" Dec 09 10:14:44 crc kubenswrapper[4824]: I1209 10:14:44.032734 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1707467c-0f4f-42a8-b964-910220950c75" path="/var/lib/kubelet/pods/1707467c-0f4f-42a8-b964-910220950c75/volumes" Dec 09 10:14:44 crc kubenswrapper[4824]: I1209 10:14:44.106541 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\") " pod="openstack/nova-metadata-0" Dec 09 10:14:44 crc kubenswrapper[4824]: I1209 10:14:44.106602 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-logs\") pod \"nova-metadata-0\" (UID: \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\") " pod="openstack/nova-metadata-0" Dec 09 10:14:44 crc kubenswrapper[4824]: I1209 10:14:44.106633 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ck8j9\" (UniqueName: \"kubernetes.io/projected/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-kube-api-access-ck8j9\") pod \"nova-metadata-0\" (UID: \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\") " pod="openstack/nova-metadata-0" Dec 09 10:14:44 crc kubenswrapper[4824]: I1209 10:14:44.106804 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-config-data\") pod \"nova-metadata-0\" (UID: \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\") " pod="openstack/nova-metadata-0" Dec 09 10:14:44 crc kubenswrapper[4824]: I1209 10:14:44.106858 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\") " pod="openstack/nova-metadata-0" Dec 09 10:14:44 crc kubenswrapper[4824]: I1209 10:14:44.107689 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-logs\") pod \"nova-metadata-0\" (UID: \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\") " pod="openstack/nova-metadata-0" Dec 09 10:14:44 crc kubenswrapper[4824]: I1209 10:14:44.118468 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\") " pod="openstack/nova-metadata-0" Dec 09 10:14:44 crc kubenswrapper[4824]: I1209 10:14:44.125523 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-config-data\") pod \"nova-metadata-0\" (UID: \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\") " pod="openstack/nova-metadata-0" Dec 09 10:14:44 crc kubenswrapper[4824]: I1209 10:14:44.127313 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\") " pod="openstack/nova-metadata-0" Dec 09 10:14:44 crc kubenswrapper[4824]: I1209 10:14:44.178061 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ck8j9\" (UniqueName: \"kubernetes.io/projected/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-kube-api-access-ck8j9\") pod \"nova-metadata-0\" (UID: \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\") " pod="openstack/nova-metadata-0" Dec 09 10:14:44 crc kubenswrapper[4824]: I1209 10:14:44.201359 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 10:14:44 crc kubenswrapper[4824]: I1209 10:14:44.849216 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 10:14:45 crc kubenswrapper[4824]: I1209 10:14:45.502825 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 09 10:14:45 crc kubenswrapper[4824]: I1209 10:14:45.553440 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 09 10:14:45 crc kubenswrapper[4824]: I1209 10:14:45.557251 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 09 10:14:45 crc kubenswrapper[4824]: I1209 10:14:45.557293 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 09 10:14:46 crc kubenswrapper[4824]: I1209 10:14:46.526689 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 09 10:14:46 crc kubenswrapper[4824]: I1209 10:14:46.638965 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9843016c-68f5-4f7a-b2a5-fe222d0ab28e" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.240:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 10:14:46 crc kubenswrapper[4824]: I1209 10:14:46.639635 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9843016c-68f5-4f7a-b2a5-fe222d0ab28e" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.240:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 10:14:47 crc kubenswrapper[4824]: I1209 10:14:47.537382 4824 generic.go:334] "Generic (PLEG): container finished" podID="de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32" containerID="a263b2169fe43a5e644746c7440339c9006ef8de764e6dcbc5e1b876e301a6c7" exitCode=0 Dec 09 10:14:47 crc kubenswrapper[4824]: I1209 10:14:47.537775 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-d6jn9" event={"ID":"de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32","Type":"ContainerDied","Data":"a263b2169fe43a5e644746c7440339c9006ef8de764e6dcbc5e1b876e301a6c7"} Dec 09 10:14:48 crc kubenswrapper[4824]: I1209 10:14:48.553376 4824 generic.go:334] "Generic (PLEG): container finished" podID="b5f03343-b401-422a-bea2-7c97f7a99600" containerID="2c46216c2b5f4579d9654864a4f17ec06cebbdf5bcccad4f0d54c87f26e021c3" exitCode=0 Dec 09 10:14:48 crc kubenswrapper[4824]: I1209 10:14:48.553455 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rcv9n" event={"ID":"b5f03343-b401-422a-bea2-7c97f7a99600","Type":"ContainerDied","Data":"2c46216c2b5f4579d9654864a4f17ec06cebbdf5bcccad4f0d54c87f26e021c3"} Dec 09 10:14:49 crc kubenswrapper[4824]: W1209 10:14:49.377582 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34aa45b3_fc9f_41f2_8c79_ee431a9f5333.slice/crio-77fed5f21696b2796dc57a8cf0a6fe5122670a3f4d1d5b7efa3d7adbf5a85331 WatchSource:0}: Error finding container 77fed5f21696b2796dc57a8cf0a6fe5122670a3f4d1d5b7efa3d7adbf5a85331: Status 404 returned error can't find the container with id 77fed5f21696b2796dc57a8cf0a6fe5122670a3f4d1d5b7efa3d7adbf5a85331 Dec 09 10:14:49 crc kubenswrapper[4824]: I1209 10:14:49.603932 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"34aa45b3-fc9f-41f2-8c79-ee431a9f5333","Type":"ContainerStarted","Data":"77fed5f21696b2796dc57a8cf0a6fe5122670a3f4d1d5b7efa3d7adbf5a85331"} Dec 09 10:14:49 crc kubenswrapper[4824]: I1209 10:14:49.610538 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-d6jn9" event={"ID":"de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32","Type":"ContainerDied","Data":"dd7f2f3f7c6c62441916fd2b9354605ee1bacf3c8f3c8b26dff5e21d52ac5c05"} Dec 09 10:14:49 crc kubenswrapper[4824]: I1209 10:14:49.610587 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd7f2f3f7c6c62441916fd2b9354605ee1bacf3c8f3c8b26dff5e21d52ac5c05" Dec 09 10:14:49 crc kubenswrapper[4824]: I1209 10:14:49.692115 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-d6jn9" Dec 09 10:14:49 crc kubenswrapper[4824]: I1209 10:14:49.802682 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32-combined-ca-bundle\") pod \"de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32\" (UID: \"de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32\") " Dec 09 10:14:49 crc kubenswrapper[4824]: I1209 10:14:49.802878 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32-config-data\") pod \"de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32\" (UID: \"de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32\") " Dec 09 10:14:49 crc kubenswrapper[4824]: I1209 10:14:49.803025 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6t7m\" (UniqueName: \"kubernetes.io/projected/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32-kube-api-access-x6t7m\") pod \"de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32\" (UID: \"de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32\") " Dec 09 10:14:49 crc kubenswrapper[4824]: I1209 10:14:49.803111 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32-scripts\") pod \"de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32\" (UID: \"de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32\") " Dec 09 10:14:49 crc kubenswrapper[4824]: I1209 10:14:49.808355 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32-kube-api-access-x6t7m" (OuterVolumeSpecName: "kube-api-access-x6t7m") pod "de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32" (UID: "de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32"). InnerVolumeSpecName "kube-api-access-x6t7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:14:49 crc kubenswrapper[4824]: I1209 10:14:49.842090 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32-scripts" (OuterVolumeSpecName: "scripts") pod "de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32" (UID: "de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:49 crc kubenswrapper[4824]: I1209 10:14:49.848343 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32" (UID: "de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:49 crc kubenswrapper[4824]: I1209 10:14:49.906976 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6t7m\" (UniqueName: \"kubernetes.io/projected/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32-kube-api-access-x6t7m\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:49 crc kubenswrapper[4824]: I1209 10:14:49.907312 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:49 crc kubenswrapper[4824]: I1209 10:14:49.907326 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:49 crc kubenswrapper[4824]: I1209 10:14:49.921388 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32-config-data" (OuterVolumeSpecName: "config-data") pod "de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32" (UID: "de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.010932 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.038746 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rcv9n" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.219062 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5f03343-b401-422a-bea2-7c97f7a99600-combined-ca-bundle\") pod \"b5f03343-b401-422a-bea2-7c97f7a99600\" (UID: \"b5f03343-b401-422a-bea2-7c97f7a99600\") " Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.219349 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7x4d\" (UniqueName: \"kubernetes.io/projected/b5f03343-b401-422a-bea2-7c97f7a99600-kube-api-access-c7x4d\") pod \"b5f03343-b401-422a-bea2-7c97f7a99600\" (UID: \"b5f03343-b401-422a-bea2-7c97f7a99600\") " Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.219394 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5f03343-b401-422a-bea2-7c97f7a99600-scripts\") pod \"b5f03343-b401-422a-bea2-7c97f7a99600\" (UID: \"b5f03343-b401-422a-bea2-7c97f7a99600\") " Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.219568 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5f03343-b401-422a-bea2-7c97f7a99600-config-data\") pod \"b5f03343-b401-422a-bea2-7c97f7a99600\" (UID: \"b5f03343-b401-422a-bea2-7c97f7a99600\") " Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.223197 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5f03343-b401-422a-bea2-7c97f7a99600-scripts" (OuterVolumeSpecName: "scripts") pod "b5f03343-b401-422a-bea2-7c97f7a99600" (UID: "b5f03343-b401-422a-bea2-7c97f7a99600"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.223696 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5f03343-b401-422a-bea2-7c97f7a99600-kube-api-access-c7x4d" (OuterVolumeSpecName: "kube-api-access-c7x4d") pod "b5f03343-b401-422a-bea2-7c97f7a99600" (UID: "b5f03343-b401-422a-bea2-7c97f7a99600"). InnerVolumeSpecName "kube-api-access-c7x4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.253757 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5f03343-b401-422a-bea2-7c97f7a99600-config-data" (OuterVolumeSpecName: "config-data") pod "b5f03343-b401-422a-bea2-7c97f7a99600" (UID: "b5f03343-b401-422a-bea2-7c97f7a99600"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.256520 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5f03343-b401-422a-bea2-7c97f7a99600-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b5f03343-b401-422a-bea2-7c97f7a99600" (UID: "b5f03343-b401-422a-bea2-7c97f7a99600"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.324083 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7x4d\" (UniqueName: \"kubernetes.io/projected/b5f03343-b401-422a-bea2-7c97f7a99600-kube-api-access-c7x4d\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.324129 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5f03343-b401-422a-bea2-7c97f7a99600-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.324140 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5f03343-b401-422a-bea2-7c97f7a99600-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.324152 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5f03343-b401-422a-bea2-7c97f7a99600-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.625942 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rcv9n" event={"ID":"b5f03343-b401-422a-bea2-7c97f7a99600","Type":"ContainerDied","Data":"4963e37b81088e0503441fed1cac51f16773707d7ec7f43b13074e3262d3cd1b"} Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.626302 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4963e37b81088e0503441fed1cac51f16773707d7ec7f43b13074e3262d3cd1b" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.626045 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rcv9n" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.628907 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"34aa45b3-fc9f-41f2-8c79-ee431a9f5333","Type":"ContainerStarted","Data":"7344da1455d6c53e1a3313feab4381426f620abf74bca32daa2e76274807042e"} Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.628984 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"34aa45b3-fc9f-41f2-8c79-ee431a9f5333","Type":"ContainerStarted","Data":"15a9018fa3a2fe941f2d971726f4aa618c698ab32f29ee6faa21c11bb8eb02b3"} Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.635874 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-d6jn9" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.636004 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-jvpcc" event={"ID":"4eb5fcc0-0113-4709-881b-7d109571a08d","Type":"ContainerStarted","Data":"5fe76dc4f6bef87e4744e566fce2c7d463ed227e3cbca0a8f76da631b8063ee1"} Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.663009 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=7.662986579 podStartE2EDuration="7.662986579s" podCreationTimestamp="2025-12-09 10:14:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:14:50.651918009 +0000 UTC m=+1646.986422676" watchObservedRunningTime="2025-12-09 10:14:50.662986579 +0000 UTC m=+1646.997491246" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.687842 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-jvpcc" podStartSLOduration=7.294701537 podStartE2EDuration="18.687819833s" podCreationTimestamp="2025-12-09 10:14:32 +0000 UTC" firstStartedPulling="2025-12-09 10:14:38.071102405 +0000 UTC m=+1634.405607072" lastFinishedPulling="2025-12-09 10:14:49.464220701 +0000 UTC m=+1645.798725368" observedRunningTime="2025-12-09 10:14:50.683126025 +0000 UTC m=+1647.017630692" watchObservedRunningTime="2025-12-09 10:14:50.687819833 +0000 UTC m=+1647.022324510" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.714943 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 09 10:14:50 crc kubenswrapper[4824]: E1209 10:14:50.715657 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32" containerName="nova-manage" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.715682 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32" containerName="nova-manage" Dec 09 10:14:50 crc kubenswrapper[4824]: E1209 10:14:50.715698 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5f03343-b401-422a-bea2-7c97f7a99600" containerName="nova-cell1-conductor-db-sync" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.715708 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5f03343-b401-422a-bea2-7c97f7a99600" containerName="nova-cell1-conductor-db-sync" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.716102 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5f03343-b401-422a-bea2-7c97f7a99600" containerName="nova-cell1-conductor-db-sync" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.716142 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32" containerName="nova-manage" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.717252 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.719729 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.759885 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.836981 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2728556-b7df-4f3b-b3d7-ec6f2683b0fa-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"a2728556-b7df-4f3b-b3d7-ec6f2683b0fa\") " pod="openstack/nova-cell1-conductor-0" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.837080 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z98xq\" (UniqueName: \"kubernetes.io/projected/a2728556-b7df-4f3b-b3d7-ec6f2683b0fa-kube-api-access-z98xq\") pod \"nova-cell1-conductor-0\" (UID: \"a2728556-b7df-4f3b-b3d7-ec6f2683b0fa\") " pod="openstack/nova-cell1-conductor-0" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.837236 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2728556-b7df-4f3b-b3d7-ec6f2683b0fa-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"a2728556-b7df-4f3b-b3d7-ec6f2683b0fa\") " pod="openstack/nova-cell1-conductor-0" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.882696 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.882986 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9843016c-68f5-4f7a-b2a5-fe222d0ab28e" containerName="nova-api-log" containerID="cri-o://07ff2b0704a6589199b5b0d22f647e6a6a4a5fa97bbc72b662fd861ededc4fad" gracePeriod=30 Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.883020 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9843016c-68f5-4f7a-b2a5-fe222d0ab28e" containerName="nova-api-api" containerID="cri-o://2fab16a089078981ace99988cb78583d5c3f2eaa3a9b19c76e2869be5cf6ee15" gracePeriod=30 Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.903087 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.903405 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="0d38851d-cb31-4701-872e-ff5d8baa1c42" containerName="nova-scheduler-scheduler" containerID="cri-o://0c0c7ca9f25255d4eb200c86bc0d5898d5bda84a3dd50cd6091a12920be7f548" gracePeriod=30 Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.924970 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.939236 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2728556-b7df-4f3b-b3d7-ec6f2683b0fa-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"a2728556-b7df-4f3b-b3d7-ec6f2683b0fa\") " pod="openstack/nova-cell1-conductor-0" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.939327 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2728556-b7df-4f3b-b3d7-ec6f2683b0fa-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"a2728556-b7df-4f3b-b3d7-ec6f2683b0fa\") " pod="openstack/nova-cell1-conductor-0" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.939399 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z98xq\" (UniqueName: \"kubernetes.io/projected/a2728556-b7df-4f3b-b3d7-ec6f2683b0fa-kube-api-access-z98xq\") pod \"nova-cell1-conductor-0\" (UID: \"a2728556-b7df-4f3b-b3d7-ec6f2683b0fa\") " pod="openstack/nova-cell1-conductor-0" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.944006 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2728556-b7df-4f3b-b3d7-ec6f2683b0fa-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"a2728556-b7df-4f3b-b3d7-ec6f2683b0fa\") " pod="openstack/nova-cell1-conductor-0" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.944446 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2728556-b7df-4f3b-b3d7-ec6f2683b0fa-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"a2728556-b7df-4f3b-b3d7-ec6f2683b0fa\") " pod="openstack/nova-cell1-conductor-0" Dec 09 10:14:50 crc kubenswrapper[4824]: I1209 10:14:50.964117 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z98xq\" (UniqueName: \"kubernetes.io/projected/a2728556-b7df-4f3b-b3d7-ec6f2683b0fa-kube-api-access-z98xq\") pod \"nova-cell1-conductor-0\" (UID: \"a2728556-b7df-4f3b-b3d7-ec6f2683b0fa\") " pod="openstack/nova-cell1-conductor-0" Dec 09 10:14:51 crc kubenswrapper[4824]: I1209 10:14:51.035125 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 09 10:14:51 crc kubenswrapper[4824]: I1209 10:14:51.577179 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 09 10:14:51 crc kubenswrapper[4824]: I1209 10:14:51.651804 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"a2728556-b7df-4f3b-b3d7-ec6f2683b0fa","Type":"ContainerStarted","Data":"05f06819fb81fc11e6f9b7cc4e3ee28cb31086e989b94d7a7bea2b13ef6a2aa1"} Dec 09 10:14:51 crc kubenswrapper[4824]: I1209 10:14:51.653698 4824 generic.go:334] "Generic (PLEG): container finished" podID="9843016c-68f5-4f7a-b2a5-fe222d0ab28e" containerID="07ff2b0704a6589199b5b0d22f647e6a6a4a5fa97bbc72b662fd861ededc4fad" exitCode=143 Dec 09 10:14:51 crc kubenswrapper[4824]: I1209 10:14:51.654732 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9843016c-68f5-4f7a-b2a5-fe222d0ab28e","Type":"ContainerDied","Data":"07ff2b0704a6589199b5b0d22f647e6a6a4a5fa97bbc72b662fd861ededc4fad"} Dec 09 10:14:52 crc kubenswrapper[4824]: I1209 10:14:52.666255 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"a2728556-b7df-4f3b-b3d7-ec6f2683b0fa","Type":"ContainerStarted","Data":"c0eb37cef0b58591f6af87129e91102fe49b72ae083286f9a902b722009e91ab"} Dec 09 10:14:52 crc kubenswrapper[4824]: I1209 10:14:52.667580 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 09 10:14:52 crc kubenswrapper[4824]: I1209 10:14:52.666772 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="34aa45b3-fc9f-41f2-8c79-ee431a9f5333" containerName="nova-metadata-metadata" containerID="cri-o://7344da1455d6c53e1a3313feab4381426f620abf74bca32daa2e76274807042e" gracePeriod=30 Dec 09 10:14:52 crc kubenswrapper[4824]: I1209 10:14:52.666424 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="34aa45b3-fc9f-41f2-8c79-ee431a9f5333" containerName="nova-metadata-log" containerID="cri-o://15a9018fa3a2fe941f2d971726f4aa618c698ab32f29ee6faa21c11bb8eb02b3" gracePeriod=30 Dec 09 10:14:52 crc kubenswrapper[4824]: I1209 10:14:52.696870 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.6968502020000003 podStartE2EDuration="2.696850202s" podCreationTimestamp="2025-12-09 10:14:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:14:52.688914992 +0000 UTC m=+1649.023419659" watchObservedRunningTime="2025-12-09 10:14:52.696850202 +0000 UTC m=+1649.031354869" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.332102 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.430535 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-config-data\") pod \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\" (UID: \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\") " Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.431139 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-combined-ca-bundle\") pod \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\" (UID: \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\") " Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.431188 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ck8j9\" (UniqueName: \"kubernetes.io/projected/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-kube-api-access-ck8j9\") pod \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\" (UID: \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\") " Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.431229 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-logs\") pod \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\" (UID: \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\") " Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.431255 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-nova-metadata-tls-certs\") pod \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\" (UID: \"34aa45b3-fc9f-41f2-8c79-ee431a9f5333\") " Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.432074 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-logs" (OuterVolumeSpecName: "logs") pod "34aa45b3-fc9f-41f2-8c79-ee431a9f5333" (UID: "34aa45b3-fc9f-41f2-8c79-ee431a9f5333"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.436139 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-kube-api-access-ck8j9" (OuterVolumeSpecName: "kube-api-access-ck8j9") pod "34aa45b3-fc9f-41f2-8c79-ee431a9f5333" (UID: "34aa45b3-fc9f-41f2-8c79-ee431a9f5333"). InnerVolumeSpecName "kube-api-access-ck8j9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.463481 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "34aa45b3-fc9f-41f2-8c79-ee431a9f5333" (UID: "34aa45b3-fc9f-41f2-8c79-ee431a9f5333"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.475812 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-config-data" (OuterVolumeSpecName: "config-data") pod "34aa45b3-fc9f-41f2-8c79-ee431a9f5333" (UID: "34aa45b3-fc9f-41f2-8c79-ee431a9f5333"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.495749 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "34aa45b3-fc9f-41f2-8c79-ee431a9f5333" (UID: "34aa45b3-fc9f-41f2-8c79-ee431a9f5333"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.535925 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.535983 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.535999 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ck8j9\" (UniqueName: \"kubernetes.io/projected/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-kube-api-access-ck8j9\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.536010 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-logs\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.536021 4824 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/34aa45b3-fc9f-41f2-8c79-ee431a9f5333-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.677313 4824 generic.go:334] "Generic (PLEG): container finished" podID="34aa45b3-fc9f-41f2-8c79-ee431a9f5333" containerID="7344da1455d6c53e1a3313feab4381426f620abf74bca32daa2e76274807042e" exitCode=0 Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.677361 4824 generic.go:334] "Generic (PLEG): container finished" podID="34aa45b3-fc9f-41f2-8c79-ee431a9f5333" containerID="15a9018fa3a2fe941f2d971726f4aa618c698ab32f29ee6faa21c11bb8eb02b3" exitCode=143 Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.677392 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.677406 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"34aa45b3-fc9f-41f2-8c79-ee431a9f5333","Type":"ContainerDied","Data":"7344da1455d6c53e1a3313feab4381426f620abf74bca32daa2e76274807042e"} Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.677445 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"34aa45b3-fc9f-41f2-8c79-ee431a9f5333","Type":"ContainerDied","Data":"15a9018fa3a2fe941f2d971726f4aa618c698ab32f29ee6faa21c11bb8eb02b3"} Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.677458 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"34aa45b3-fc9f-41f2-8c79-ee431a9f5333","Type":"ContainerDied","Data":"77fed5f21696b2796dc57a8cf0a6fe5122670a3f4d1d5b7efa3d7adbf5a85331"} Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.677475 4824 scope.go:117] "RemoveContainer" containerID="7344da1455d6c53e1a3313feab4381426f620abf74bca32daa2e76274807042e" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.681117 4824 generic.go:334] "Generic (PLEG): container finished" podID="4eb5fcc0-0113-4709-881b-7d109571a08d" containerID="5fe76dc4f6bef87e4744e566fce2c7d463ed227e3cbca0a8f76da631b8063ee1" exitCode=0 Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.682249 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-jvpcc" event={"ID":"4eb5fcc0-0113-4709-881b-7d109571a08d","Type":"ContainerDied","Data":"5fe76dc4f6bef87e4744e566fce2c7d463ed227e3cbca0a8f76da631b8063ee1"} Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.688009 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.767603 4824 scope.go:117] "RemoveContainer" containerID="15a9018fa3a2fe941f2d971726f4aa618c698ab32f29ee6faa21c11bb8eb02b3" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.815067 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.830499 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.839035 4824 scope.go:117] "RemoveContainer" containerID="7344da1455d6c53e1a3313feab4381426f620abf74bca32daa2e76274807042e" Dec 09 10:14:53 crc kubenswrapper[4824]: E1209 10:14:53.839945 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7344da1455d6c53e1a3313feab4381426f620abf74bca32daa2e76274807042e\": container with ID starting with 7344da1455d6c53e1a3313feab4381426f620abf74bca32daa2e76274807042e not found: ID does not exist" containerID="7344da1455d6c53e1a3313feab4381426f620abf74bca32daa2e76274807042e" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.840006 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7344da1455d6c53e1a3313feab4381426f620abf74bca32daa2e76274807042e"} err="failed to get container status \"7344da1455d6c53e1a3313feab4381426f620abf74bca32daa2e76274807042e\": rpc error: code = NotFound desc = could not find container \"7344da1455d6c53e1a3313feab4381426f620abf74bca32daa2e76274807042e\": container with ID starting with 7344da1455d6c53e1a3313feab4381426f620abf74bca32daa2e76274807042e not found: ID does not exist" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.840040 4824 scope.go:117] "RemoveContainer" containerID="15a9018fa3a2fe941f2d971726f4aa618c698ab32f29ee6faa21c11bb8eb02b3" Dec 09 10:14:53 crc kubenswrapper[4824]: E1209 10:14:53.840406 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15a9018fa3a2fe941f2d971726f4aa618c698ab32f29ee6faa21c11bb8eb02b3\": container with ID starting with 15a9018fa3a2fe941f2d971726f4aa618c698ab32f29ee6faa21c11bb8eb02b3 not found: ID does not exist" containerID="15a9018fa3a2fe941f2d971726f4aa618c698ab32f29ee6faa21c11bb8eb02b3" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.840499 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15a9018fa3a2fe941f2d971726f4aa618c698ab32f29ee6faa21c11bb8eb02b3"} err="failed to get container status \"15a9018fa3a2fe941f2d971726f4aa618c698ab32f29ee6faa21c11bb8eb02b3\": rpc error: code = NotFound desc = could not find container \"15a9018fa3a2fe941f2d971726f4aa618c698ab32f29ee6faa21c11bb8eb02b3\": container with ID starting with 15a9018fa3a2fe941f2d971726f4aa618c698ab32f29ee6faa21c11bb8eb02b3 not found: ID does not exist" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.840584 4824 scope.go:117] "RemoveContainer" containerID="7344da1455d6c53e1a3313feab4381426f620abf74bca32daa2e76274807042e" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.841636 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7344da1455d6c53e1a3313feab4381426f620abf74bca32daa2e76274807042e"} err="failed to get container status \"7344da1455d6c53e1a3313feab4381426f620abf74bca32daa2e76274807042e\": rpc error: code = NotFound desc = could not find container \"7344da1455d6c53e1a3313feab4381426f620abf74bca32daa2e76274807042e\": container with ID starting with 7344da1455d6c53e1a3313feab4381426f620abf74bca32daa2e76274807042e not found: ID does not exist" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.841681 4824 scope.go:117] "RemoveContainer" containerID="15a9018fa3a2fe941f2d971726f4aa618c698ab32f29ee6faa21c11bb8eb02b3" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.842022 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15a9018fa3a2fe941f2d971726f4aa618c698ab32f29ee6faa21c11bb8eb02b3"} err="failed to get container status \"15a9018fa3a2fe941f2d971726f4aa618c698ab32f29ee6faa21c11bb8eb02b3\": rpc error: code = NotFound desc = could not find container \"15a9018fa3a2fe941f2d971726f4aa618c698ab32f29ee6faa21c11bb8eb02b3\": container with ID starting with 15a9018fa3a2fe941f2d971726f4aa618c698ab32f29ee6faa21c11bb8eb02b3 not found: ID does not exist" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.847275 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 09 10:14:53 crc kubenswrapper[4824]: E1209 10:14:53.847996 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34aa45b3-fc9f-41f2-8c79-ee431a9f5333" containerName="nova-metadata-metadata" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.848019 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="34aa45b3-fc9f-41f2-8c79-ee431a9f5333" containerName="nova-metadata-metadata" Dec 09 10:14:53 crc kubenswrapper[4824]: E1209 10:14:53.848056 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34aa45b3-fc9f-41f2-8c79-ee431a9f5333" containerName="nova-metadata-log" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.848064 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="34aa45b3-fc9f-41f2-8c79-ee431a9f5333" containerName="nova-metadata-log" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.848340 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="34aa45b3-fc9f-41f2-8c79-ee431a9f5333" containerName="nova-metadata-metadata" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.848370 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="34aa45b3-fc9f-41f2-8c79-ee431a9f5333" containerName="nova-metadata-log" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.850043 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.855954 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.856161 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.873021 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.932630 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34aa45b3-fc9f-41f2-8c79-ee431a9f5333" path="/var/lib/kubelet/pods/34aa45b3-fc9f-41f2-8c79-ee431a9f5333/volumes" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.948194 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\") " pod="openstack/nova-metadata-0" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.948542 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4rwc\" (UniqueName: \"kubernetes.io/projected/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-kube-api-access-m4rwc\") pod \"nova-metadata-0\" (UID: \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\") " pod="openstack/nova-metadata-0" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.948725 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-config-data\") pod \"nova-metadata-0\" (UID: \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\") " pod="openstack/nova-metadata-0" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.948839 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-logs\") pod \"nova-metadata-0\" (UID: \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\") " pod="openstack/nova-metadata-0" Dec 09 10:14:53 crc kubenswrapper[4824]: I1209 10:14:53.948926 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\") " pod="openstack/nova-metadata-0" Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.051832 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\") " pod="openstack/nova-metadata-0" Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.052212 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4rwc\" (UniqueName: \"kubernetes.io/projected/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-kube-api-access-m4rwc\") pod \"nova-metadata-0\" (UID: \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\") " pod="openstack/nova-metadata-0" Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.052364 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-config-data\") pod \"nova-metadata-0\" (UID: \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\") " pod="openstack/nova-metadata-0" Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.052406 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-logs\") pod \"nova-metadata-0\" (UID: \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\") " pod="openstack/nova-metadata-0" Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.052452 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\") " pod="openstack/nova-metadata-0" Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.054182 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-logs\") pod \"nova-metadata-0\" (UID: \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\") " pod="openstack/nova-metadata-0" Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.057622 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-config-data\") pod \"nova-metadata-0\" (UID: \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\") " pod="openstack/nova-metadata-0" Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.058859 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\") " pod="openstack/nova-metadata-0" Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.058961 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\") " pod="openstack/nova-metadata-0" Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.075333 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4rwc\" (UniqueName: \"kubernetes.io/projected/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-kube-api-access-m4rwc\") pod \"nova-metadata-0\" (UID: \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\") " pod="openstack/nova-metadata-0" Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.231841 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.489075 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.675713 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9843016c-68f5-4f7a-b2a5-fe222d0ab28e-logs\") pod \"9843016c-68f5-4f7a-b2a5-fe222d0ab28e\" (UID: \"9843016c-68f5-4f7a-b2a5-fe222d0ab28e\") " Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.675940 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9843016c-68f5-4f7a-b2a5-fe222d0ab28e-combined-ca-bundle\") pod \"9843016c-68f5-4f7a-b2a5-fe222d0ab28e\" (UID: \"9843016c-68f5-4f7a-b2a5-fe222d0ab28e\") " Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.676015 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hf2j\" (UniqueName: \"kubernetes.io/projected/9843016c-68f5-4f7a-b2a5-fe222d0ab28e-kube-api-access-8hf2j\") pod \"9843016c-68f5-4f7a-b2a5-fe222d0ab28e\" (UID: \"9843016c-68f5-4f7a-b2a5-fe222d0ab28e\") " Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.676940 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9843016c-68f5-4f7a-b2a5-fe222d0ab28e-config-data\") pod \"9843016c-68f5-4f7a-b2a5-fe222d0ab28e\" (UID: \"9843016c-68f5-4f7a-b2a5-fe222d0ab28e\") " Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.678216 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9843016c-68f5-4f7a-b2a5-fe222d0ab28e-logs" (OuterVolumeSpecName: "logs") pod "9843016c-68f5-4f7a-b2a5-fe222d0ab28e" (UID: "9843016c-68f5-4f7a-b2a5-fe222d0ab28e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.683125 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9843016c-68f5-4f7a-b2a5-fe222d0ab28e-kube-api-access-8hf2j" (OuterVolumeSpecName: "kube-api-access-8hf2j") pod "9843016c-68f5-4f7a-b2a5-fe222d0ab28e" (UID: "9843016c-68f5-4f7a-b2a5-fe222d0ab28e"). InnerVolumeSpecName "kube-api-access-8hf2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.709480 4824 generic.go:334] "Generic (PLEG): container finished" podID="9843016c-68f5-4f7a-b2a5-fe222d0ab28e" containerID="2fab16a089078981ace99988cb78583d5c3f2eaa3a9b19c76e2869be5cf6ee15" exitCode=0 Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.709565 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9843016c-68f5-4f7a-b2a5-fe222d0ab28e","Type":"ContainerDied","Data":"2fab16a089078981ace99988cb78583d5c3f2eaa3a9b19c76e2869be5cf6ee15"} Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.709616 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9843016c-68f5-4f7a-b2a5-fe222d0ab28e","Type":"ContainerDied","Data":"2308b66baa8a7f643a046ca9bb78b35b295d7d05dad622b4c8ed4f873231548a"} Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.709647 4824 scope.go:117] "RemoveContainer" containerID="2fab16a089078981ace99988cb78583d5c3f2eaa3a9b19c76e2869be5cf6ee15" Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.709890 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.848562 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9843016c-68f5-4f7a-b2a5-fe222d0ab28e-logs\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.849012 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hf2j\" (UniqueName: \"kubernetes.io/projected/9843016c-68f5-4f7a-b2a5-fe222d0ab28e-kube-api-access-8hf2j\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.852952 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9843016c-68f5-4f7a-b2a5-fe222d0ab28e-config-data" (OuterVolumeSpecName: "config-data") pod "9843016c-68f5-4f7a-b2a5-fe222d0ab28e" (UID: "9843016c-68f5-4f7a-b2a5-fe222d0ab28e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.857064 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9843016c-68f5-4f7a-b2a5-fe222d0ab28e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9843016c-68f5-4f7a-b2a5-fe222d0ab28e" (UID: "9843016c-68f5-4f7a-b2a5-fe222d0ab28e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.887886 4824 scope.go:117] "RemoveContainer" containerID="07ff2b0704a6589199b5b0d22f647e6a6a4a5fa97bbc72b662fd861ededc4fad" Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.913048 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 10:14:54 crc kubenswrapper[4824]: W1209 10:14:54.940113 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b1dd9ff_c6e7_4b2f_92c2_b0c020b5783c.slice/crio-8ac6f318e9cd03ac199b5d9a39189b8f2123d11c787b916e6485dd7a3e72e8cf WatchSource:0}: Error finding container 8ac6f318e9cd03ac199b5d9a39189b8f2123d11c787b916e6485dd7a3e72e8cf: Status 404 returned error can't find the container with id 8ac6f318e9cd03ac199b5d9a39189b8f2123d11c787b916e6485dd7a3e72e8cf Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.958528 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9843016c-68f5-4f7a-b2a5-fe222d0ab28e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:54 crc kubenswrapper[4824]: I1209 10:14:54.958571 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9843016c-68f5-4f7a-b2a5-fe222d0ab28e-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.040117 4824 scope.go:117] "RemoveContainer" containerID="2fab16a089078981ace99988cb78583d5c3f2eaa3a9b19c76e2869be5cf6ee15" Dec 09 10:14:55 crc kubenswrapper[4824]: E1209 10:14:55.040532 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fab16a089078981ace99988cb78583d5c3f2eaa3a9b19c76e2869be5cf6ee15\": container with ID starting with 2fab16a089078981ace99988cb78583d5c3f2eaa3a9b19c76e2869be5cf6ee15 not found: ID does not exist" containerID="2fab16a089078981ace99988cb78583d5c3f2eaa3a9b19c76e2869be5cf6ee15" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.040574 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fab16a089078981ace99988cb78583d5c3f2eaa3a9b19c76e2869be5cf6ee15"} err="failed to get container status \"2fab16a089078981ace99988cb78583d5c3f2eaa3a9b19c76e2869be5cf6ee15\": rpc error: code = NotFound desc = could not find container \"2fab16a089078981ace99988cb78583d5c3f2eaa3a9b19c76e2869be5cf6ee15\": container with ID starting with 2fab16a089078981ace99988cb78583d5c3f2eaa3a9b19c76e2869be5cf6ee15 not found: ID does not exist" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.040601 4824 scope.go:117] "RemoveContainer" containerID="07ff2b0704a6589199b5b0d22f647e6a6a4a5fa97bbc72b662fd861ededc4fad" Dec 09 10:14:55 crc kubenswrapper[4824]: E1209 10:14:55.041451 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07ff2b0704a6589199b5b0d22f647e6a6a4a5fa97bbc72b662fd861ededc4fad\": container with ID starting with 07ff2b0704a6589199b5b0d22f647e6a6a4a5fa97bbc72b662fd861ededc4fad not found: ID does not exist" containerID="07ff2b0704a6589199b5b0d22f647e6a6a4a5fa97bbc72b662fd861ededc4fad" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.041513 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07ff2b0704a6589199b5b0d22f647e6a6a4a5fa97bbc72b662fd861ededc4fad"} err="failed to get container status \"07ff2b0704a6589199b5b0d22f647e6a6a4a5fa97bbc72b662fd861ededc4fad\": rpc error: code = NotFound desc = could not find container \"07ff2b0704a6589199b5b0d22f647e6a6a4a5fa97bbc72b662fd861ededc4fad\": container with ID starting with 07ff2b0704a6589199b5b0d22f647e6a6a4a5fa97bbc72b662fd861ededc4fad not found: ID does not exist" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.180295 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.215402 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.236773 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 09 10:14:55 crc kubenswrapper[4824]: E1209 10:14:55.237761 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9843016c-68f5-4f7a-b2a5-fe222d0ab28e" containerName="nova-api-log" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.237802 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9843016c-68f5-4f7a-b2a5-fe222d0ab28e" containerName="nova-api-log" Dec 09 10:14:55 crc kubenswrapper[4824]: E1209 10:14:55.237831 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9843016c-68f5-4f7a-b2a5-fe222d0ab28e" containerName="nova-api-api" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.237837 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9843016c-68f5-4f7a-b2a5-fe222d0ab28e" containerName="nova-api-api" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.238314 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9843016c-68f5-4f7a-b2a5-fe222d0ab28e" containerName="nova-api-api" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.238366 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9843016c-68f5-4f7a-b2a5-fe222d0ab28e" containerName="nova-api-log" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.244247 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.252769 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.257225 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.286634 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lgbh\" (UniqueName: \"kubernetes.io/projected/74053563-b115-4fb2-9433-16e81288df8d-kube-api-access-4lgbh\") pod \"nova-api-0\" (UID: \"74053563-b115-4fb2-9433-16e81288df8d\") " pod="openstack/nova-api-0" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.286690 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74053563-b115-4fb2-9433-16e81288df8d-config-data\") pod \"nova-api-0\" (UID: \"74053563-b115-4fb2-9433-16e81288df8d\") " pod="openstack/nova-api-0" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.287031 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74053563-b115-4fb2-9433-16e81288df8d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"74053563-b115-4fb2-9433-16e81288df8d\") " pod="openstack/nova-api-0" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.287255 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74053563-b115-4fb2-9433-16e81288df8d-logs\") pod \"nova-api-0\" (UID: \"74053563-b115-4fb2-9433-16e81288df8d\") " pod="openstack/nova-api-0" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.378600 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-jvpcc" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.391164 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74053563-b115-4fb2-9433-16e81288df8d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"74053563-b115-4fb2-9433-16e81288df8d\") " pod="openstack/nova-api-0" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.391307 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74053563-b115-4fb2-9433-16e81288df8d-logs\") pod \"nova-api-0\" (UID: \"74053563-b115-4fb2-9433-16e81288df8d\") " pod="openstack/nova-api-0" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.391446 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lgbh\" (UniqueName: \"kubernetes.io/projected/74053563-b115-4fb2-9433-16e81288df8d-kube-api-access-4lgbh\") pod \"nova-api-0\" (UID: \"74053563-b115-4fb2-9433-16e81288df8d\") " pod="openstack/nova-api-0" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.391476 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74053563-b115-4fb2-9433-16e81288df8d-config-data\") pod \"nova-api-0\" (UID: \"74053563-b115-4fb2-9433-16e81288df8d\") " pod="openstack/nova-api-0" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.393028 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74053563-b115-4fb2-9433-16e81288df8d-logs\") pod \"nova-api-0\" (UID: \"74053563-b115-4fb2-9433-16e81288df8d\") " pod="openstack/nova-api-0" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.400983 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74053563-b115-4fb2-9433-16e81288df8d-config-data\") pod \"nova-api-0\" (UID: \"74053563-b115-4fb2-9433-16e81288df8d\") " pod="openstack/nova-api-0" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.401266 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74053563-b115-4fb2-9433-16e81288df8d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"74053563-b115-4fb2-9433-16e81288df8d\") " pod="openstack/nova-api-0" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.435105 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lgbh\" (UniqueName: \"kubernetes.io/projected/74053563-b115-4fb2-9433-16e81288df8d-kube-api-access-4lgbh\") pod \"nova-api-0\" (UID: \"74053563-b115-4fb2-9433-16e81288df8d\") " pod="openstack/nova-api-0" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.496134 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4eb5fcc0-0113-4709-881b-7d109571a08d-scripts\") pod \"4eb5fcc0-0113-4709-881b-7d109571a08d\" (UID: \"4eb5fcc0-0113-4709-881b-7d109571a08d\") " Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.496229 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wdph\" (UniqueName: \"kubernetes.io/projected/4eb5fcc0-0113-4709-881b-7d109571a08d-kube-api-access-6wdph\") pod \"4eb5fcc0-0113-4709-881b-7d109571a08d\" (UID: \"4eb5fcc0-0113-4709-881b-7d109571a08d\") " Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.496426 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eb5fcc0-0113-4709-881b-7d109571a08d-config-data\") pod \"4eb5fcc0-0113-4709-881b-7d109571a08d\" (UID: \"4eb5fcc0-0113-4709-881b-7d109571a08d\") " Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.496675 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eb5fcc0-0113-4709-881b-7d109571a08d-combined-ca-bundle\") pod \"4eb5fcc0-0113-4709-881b-7d109571a08d\" (UID: \"4eb5fcc0-0113-4709-881b-7d109571a08d\") " Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.508190 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eb5fcc0-0113-4709-881b-7d109571a08d-scripts" (OuterVolumeSpecName: "scripts") pod "4eb5fcc0-0113-4709-881b-7d109571a08d" (UID: "4eb5fcc0-0113-4709-881b-7d109571a08d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.528009 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eb5fcc0-0113-4709-881b-7d109571a08d-kube-api-access-6wdph" (OuterVolumeSpecName: "kube-api-access-6wdph") pod "4eb5fcc0-0113-4709-881b-7d109571a08d" (UID: "4eb5fcc0-0113-4709-881b-7d109571a08d"). InnerVolumeSpecName "kube-api-access-6wdph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:14:55 crc kubenswrapper[4824]: E1209 10:14:55.548206 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0c0c7ca9f25255d4eb200c86bc0d5898d5bda84a3dd50cd6091a12920be7f548" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 10:14:55 crc kubenswrapper[4824]: E1209 10:14:55.548692 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0c0c7ca9f25255d4eb200c86bc0d5898d5bda84a3dd50cd6091a12920be7f548 is running failed: container process not found" containerID="0c0c7ca9f25255d4eb200c86bc0d5898d5bda84a3dd50cd6091a12920be7f548" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 10:14:55 crc kubenswrapper[4824]: E1209 10:14:55.556395 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0c0c7ca9f25255d4eb200c86bc0d5898d5bda84a3dd50cd6091a12920be7f548 is running failed: container process not found" containerID="0c0c7ca9f25255d4eb200c86bc0d5898d5bda84a3dd50cd6091a12920be7f548" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 10:14:55 crc kubenswrapper[4824]: E1209 10:14:55.556468 4824 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0c0c7ca9f25255d4eb200c86bc0d5898d5bda84a3dd50cd6091a12920be7f548 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="0d38851d-cb31-4701-872e-ff5d8baa1c42" containerName="nova-scheduler-scheduler" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.594557 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eb5fcc0-0113-4709-881b-7d109571a08d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4eb5fcc0-0113-4709-881b-7d109571a08d" (UID: "4eb5fcc0-0113-4709-881b-7d109571a08d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.600449 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eb5fcc0-0113-4709-881b-7d109571a08d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.600589 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4eb5fcc0-0113-4709-881b-7d109571a08d-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.600600 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wdph\" (UniqueName: \"kubernetes.io/projected/4eb5fcc0-0113-4709-881b-7d109571a08d-kube-api-access-6wdph\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.601124 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eb5fcc0-0113-4709-881b-7d109571a08d-config-data" (OuterVolumeSpecName: "config-data") pod "4eb5fcc0-0113-4709-881b-7d109571a08d" (UID: "4eb5fcc0-0113-4709-881b-7d109571a08d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.697961 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.704509 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eb5fcc0-0113-4709-881b-7d109571a08d-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.754980 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c","Type":"ContainerStarted","Data":"3dca08d942566f4202f7339b1ab8c1a09137689ed513a884194c5c0b7c8c7b3c"} Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.755047 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c","Type":"ContainerStarted","Data":"8ac6f318e9cd03ac199b5d9a39189b8f2123d11c787b916e6485dd7a3e72e8cf"} Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.757342 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-jvpcc" event={"ID":"4eb5fcc0-0113-4709-881b-7d109571a08d","Type":"ContainerDied","Data":"f313b8129a47b7e4efe43b22e6737ee535fbba1bcc54ef995d3ac7edd41d93e4"} Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.757395 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f313b8129a47b7e4efe43b22e6737ee535fbba1bcc54ef995d3ac7edd41d93e4" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.757473 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-jvpcc" Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.765085 4824 generic.go:334] "Generic (PLEG): container finished" podID="0d38851d-cb31-4701-872e-ff5d8baa1c42" containerID="0c0c7ca9f25255d4eb200c86bc0d5898d5bda84a3dd50cd6091a12920be7f548" exitCode=0 Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.765208 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0d38851d-cb31-4701-872e-ff5d8baa1c42","Type":"ContainerDied","Data":"0c0c7ca9f25255d4eb200c86bc0d5898d5bda84a3dd50cd6091a12920be7f548"} Dec 09 10:14:55 crc kubenswrapper[4824]: I1209 10:14:55.949741 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9843016c-68f5-4f7a-b2a5-fe222d0ab28e" path="/var/lib/kubelet/pods/9843016c-68f5-4f7a-b2a5-fe222d0ab28e/volumes" Dec 09 10:14:56 crc kubenswrapper[4824]: I1209 10:14:56.080356 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 09 10:14:56 crc kubenswrapper[4824]: I1209 10:14:56.265130 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 10:14:56 crc kubenswrapper[4824]: I1209 10:14:56.328894 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d38851d-cb31-4701-872e-ff5d8baa1c42-config-data\") pod \"0d38851d-cb31-4701-872e-ff5d8baa1c42\" (UID: \"0d38851d-cb31-4701-872e-ff5d8baa1c42\") " Dec 09 10:14:56 crc kubenswrapper[4824]: I1209 10:14:56.329410 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d38851d-cb31-4701-872e-ff5d8baa1c42-combined-ca-bundle\") pod \"0d38851d-cb31-4701-872e-ff5d8baa1c42\" (UID: \"0d38851d-cb31-4701-872e-ff5d8baa1c42\") " Dec 09 10:14:56 crc kubenswrapper[4824]: I1209 10:14:56.329495 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhxd7\" (UniqueName: \"kubernetes.io/projected/0d38851d-cb31-4701-872e-ff5d8baa1c42-kube-api-access-nhxd7\") pod \"0d38851d-cb31-4701-872e-ff5d8baa1c42\" (UID: \"0d38851d-cb31-4701-872e-ff5d8baa1c42\") " Dec 09 10:14:56 crc kubenswrapper[4824]: I1209 10:14:56.336343 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d38851d-cb31-4701-872e-ff5d8baa1c42-kube-api-access-nhxd7" (OuterVolumeSpecName: "kube-api-access-nhxd7") pod "0d38851d-cb31-4701-872e-ff5d8baa1c42" (UID: "0d38851d-cb31-4701-872e-ff5d8baa1c42"). InnerVolumeSpecName "kube-api-access-nhxd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:14:56 crc kubenswrapper[4824]: I1209 10:14:56.359472 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 10:14:56 crc kubenswrapper[4824]: I1209 10:14:56.376933 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d38851d-cb31-4701-872e-ff5d8baa1c42-config-data" (OuterVolumeSpecName: "config-data") pod "0d38851d-cb31-4701-872e-ff5d8baa1c42" (UID: "0d38851d-cb31-4701-872e-ff5d8baa1c42"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:56 crc kubenswrapper[4824]: I1209 10:14:56.407329 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d38851d-cb31-4701-872e-ff5d8baa1c42-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d38851d-cb31-4701-872e-ff5d8baa1c42" (UID: "0d38851d-cb31-4701-872e-ff5d8baa1c42"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:14:56 crc kubenswrapper[4824]: I1209 10:14:56.432443 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d38851d-cb31-4701-872e-ff5d8baa1c42-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:56 crc kubenswrapper[4824]: I1209 10:14:56.432488 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhxd7\" (UniqueName: \"kubernetes.io/projected/0d38851d-cb31-4701-872e-ff5d8baa1c42-kube-api-access-nhxd7\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:56 crc kubenswrapper[4824]: I1209 10:14:56.432505 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d38851d-cb31-4701-872e-ff5d8baa1c42-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:14:56 crc kubenswrapper[4824]: I1209 10:14:56.782393 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c","Type":"ContainerStarted","Data":"8b9908e6657ec4e808f35ce0be343b0d2e3a72bfb3fb7f6d9ec725e55361459b"} Dec 09 10:14:56 crc kubenswrapper[4824]: I1209 10:14:56.789580 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"74053563-b115-4fb2-9433-16e81288df8d","Type":"ContainerStarted","Data":"4b6d495e7633ad7352a5222410c3cfc1f99d6b25ec2d79c143b79fe408a02c74"} Dec 09 10:14:56 crc kubenswrapper[4824]: I1209 10:14:56.789638 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"74053563-b115-4fb2-9433-16e81288df8d","Type":"ContainerStarted","Data":"1dd6986809d86b0375eba3e70ea42bde9ef897c1c0331e219811b0dcc6c09a73"} Dec 09 10:14:56 crc kubenswrapper[4824]: I1209 10:14:56.789660 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"74053563-b115-4fb2-9433-16e81288df8d","Type":"ContainerStarted","Data":"26dfbca4840c6bb147000031c2cd4d7b5551267e552f941277a67648603422d0"} Dec 09 10:14:56 crc kubenswrapper[4824]: I1209 10:14:56.795650 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0d38851d-cb31-4701-872e-ff5d8baa1c42","Type":"ContainerDied","Data":"6a14b58f788c990f664be3a452a265ba4d25852931272770465624184400ac35"} Dec 09 10:14:56 crc kubenswrapper[4824]: I1209 10:14:56.795718 4824 scope.go:117] "RemoveContainer" containerID="0c0c7ca9f25255d4eb200c86bc0d5898d5bda84a3dd50cd6091a12920be7f548" Dec 09 10:14:56 crc kubenswrapper[4824]: I1209 10:14:56.795821 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 10:14:56 crc kubenswrapper[4824]: I1209 10:14:56.813294 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.813276848 podStartE2EDuration="3.813276848s" podCreationTimestamp="2025-12-09 10:14:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:14:56.81047628 +0000 UTC m=+1653.144980947" watchObservedRunningTime="2025-12-09 10:14:56.813276848 +0000 UTC m=+1653.147781515" Dec 09 10:14:56 crc kubenswrapper[4824]: I1209 10:14:56.842415 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.842377618 podStartE2EDuration="1.842377618s" podCreationTimestamp="2025-12-09 10:14:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:14:56.840479978 +0000 UTC m=+1653.174984645" watchObservedRunningTime="2025-12-09 10:14:56.842377618 +0000 UTC m=+1653.176882285" Dec 09 10:14:56 crc kubenswrapper[4824]: I1209 10:14:56.956945 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 10:14:56 crc kubenswrapper[4824]: I1209 10:14:56.976307 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 10:14:56 crc kubenswrapper[4824]: I1209 10:14:56.996764 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 10:14:57 crc kubenswrapper[4824]: E1209 10:14:57.000330 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eb5fcc0-0113-4709-881b-7d109571a08d" containerName="aodh-db-sync" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.000373 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eb5fcc0-0113-4709-881b-7d109571a08d" containerName="aodh-db-sync" Dec 09 10:14:57 crc kubenswrapper[4824]: E1209 10:14:57.000443 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d38851d-cb31-4701-872e-ff5d8baa1c42" containerName="nova-scheduler-scheduler" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.000449 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d38851d-cb31-4701-872e-ff5d8baa1c42" containerName="nova-scheduler-scheduler" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.000851 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d38851d-cb31-4701-872e-ff5d8baa1c42" containerName="nova-scheduler-scheduler" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.000887 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eb5fcc0-0113-4709-881b-7d109571a08d" containerName="aodh-db-sync" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.001755 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.006239 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.014493 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.055190 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/889dfa64-048d-4d55-a734-5b179dfcc1f5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"889dfa64-048d-4d55-a734-5b179dfcc1f5\") " pod="openstack/nova-scheduler-0" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.055317 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/889dfa64-048d-4d55-a734-5b179dfcc1f5-config-data\") pod \"nova-scheduler-0\" (UID: \"889dfa64-048d-4d55-a734-5b179dfcc1f5\") " pod="openstack/nova-scheduler-0" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.055352 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m69tb\" (UniqueName: \"kubernetes.io/projected/889dfa64-048d-4d55-a734-5b179dfcc1f5-kube-api-access-m69tb\") pod \"nova-scheduler-0\" (UID: \"889dfa64-048d-4d55-a734-5b179dfcc1f5\") " pod="openstack/nova-scheduler-0" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.157769 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/889dfa64-048d-4d55-a734-5b179dfcc1f5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"889dfa64-048d-4d55-a734-5b179dfcc1f5\") " pod="openstack/nova-scheduler-0" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.157938 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/889dfa64-048d-4d55-a734-5b179dfcc1f5-config-data\") pod \"nova-scheduler-0\" (UID: \"889dfa64-048d-4d55-a734-5b179dfcc1f5\") " pod="openstack/nova-scheduler-0" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.157978 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m69tb\" (UniqueName: \"kubernetes.io/projected/889dfa64-048d-4d55-a734-5b179dfcc1f5-kube-api-access-m69tb\") pod \"nova-scheduler-0\" (UID: \"889dfa64-048d-4d55-a734-5b179dfcc1f5\") " pod="openstack/nova-scheduler-0" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.163624 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/889dfa64-048d-4d55-a734-5b179dfcc1f5-config-data\") pod \"nova-scheduler-0\" (UID: \"889dfa64-048d-4d55-a734-5b179dfcc1f5\") " pod="openstack/nova-scheduler-0" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.170630 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/889dfa64-048d-4d55-a734-5b179dfcc1f5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"889dfa64-048d-4d55-a734-5b179dfcc1f5\") " pod="openstack/nova-scheduler-0" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.178586 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m69tb\" (UniqueName: \"kubernetes.io/projected/889dfa64-048d-4d55-a734-5b179dfcc1f5-kube-api-access-m69tb\") pod \"nova-scheduler-0\" (UID: \"889dfa64-048d-4d55-a734-5b179dfcc1f5\") " pod="openstack/nova-scheduler-0" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.328721 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.706857 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.711385 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.726431 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-mgzrr" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.726951 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.726980 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.727091 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.866126 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.876859 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/384b8d7f-e60a-4421-855d-6d397385f7a8-config-data\") pod \"aodh-0\" (UID: \"384b8d7f-e60a-4421-855d-6d397385f7a8\") " pod="openstack/aodh-0" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.876983 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/384b8d7f-e60a-4421-855d-6d397385f7a8-combined-ca-bundle\") pod \"aodh-0\" (UID: \"384b8d7f-e60a-4421-855d-6d397385f7a8\") " pod="openstack/aodh-0" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.877110 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/384b8d7f-e60a-4421-855d-6d397385f7a8-scripts\") pod \"aodh-0\" (UID: \"384b8d7f-e60a-4421-855d-6d397385f7a8\") " pod="openstack/aodh-0" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.877299 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbjwc\" (UniqueName: \"kubernetes.io/projected/384b8d7f-e60a-4421-855d-6d397385f7a8-kube-api-access-qbjwc\") pod \"aodh-0\" (UID: \"384b8d7f-e60a-4421-855d-6d397385f7a8\") " pod="openstack/aodh-0" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.934244 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d38851d-cb31-4701-872e-ff5d8baa1c42" path="/var/lib/kubelet/pods/0d38851d-cb31-4701-872e-ff5d8baa1c42/volumes" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.981040 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbjwc\" (UniqueName: \"kubernetes.io/projected/384b8d7f-e60a-4421-855d-6d397385f7a8-kube-api-access-qbjwc\") pod \"aodh-0\" (UID: \"384b8d7f-e60a-4421-855d-6d397385f7a8\") " pod="openstack/aodh-0" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.981161 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/384b8d7f-e60a-4421-855d-6d397385f7a8-config-data\") pod \"aodh-0\" (UID: \"384b8d7f-e60a-4421-855d-6d397385f7a8\") " pod="openstack/aodh-0" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.981228 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/384b8d7f-e60a-4421-855d-6d397385f7a8-combined-ca-bundle\") pod \"aodh-0\" (UID: \"384b8d7f-e60a-4421-855d-6d397385f7a8\") " pod="openstack/aodh-0" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.981334 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/384b8d7f-e60a-4421-855d-6d397385f7a8-scripts\") pod \"aodh-0\" (UID: \"384b8d7f-e60a-4421-855d-6d397385f7a8\") " pod="openstack/aodh-0" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.991582 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/384b8d7f-e60a-4421-855d-6d397385f7a8-combined-ca-bundle\") pod \"aodh-0\" (UID: \"384b8d7f-e60a-4421-855d-6d397385f7a8\") " pod="openstack/aodh-0" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.991971 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/384b8d7f-e60a-4421-855d-6d397385f7a8-scripts\") pod \"aodh-0\" (UID: \"384b8d7f-e60a-4421-855d-6d397385f7a8\") " pod="openstack/aodh-0" Dec 09 10:14:57 crc kubenswrapper[4824]: I1209 10:14:57.992629 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/384b8d7f-e60a-4421-855d-6d397385f7a8-config-data\") pod \"aodh-0\" (UID: \"384b8d7f-e60a-4421-855d-6d397385f7a8\") " pod="openstack/aodh-0" Dec 09 10:14:58 crc kubenswrapper[4824]: I1209 10:14:58.014892 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbjwc\" (UniqueName: \"kubernetes.io/projected/384b8d7f-e60a-4421-855d-6d397385f7a8-kube-api-access-qbjwc\") pod \"aodh-0\" (UID: \"384b8d7f-e60a-4421-855d-6d397385f7a8\") " pod="openstack/aodh-0" Dec 09 10:14:58 crc kubenswrapper[4824]: I1209 10:14:58.051362 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 09 10:14:58 crc kubenswrapper[4824]: I1209 10:14:58.692608 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 09 10:14:58 crc kubenswrapper[4824]: W1209 10:14:58.694202 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod384b8d7f_e60a_4421_855d_6d397385f7a8.slice/crio-810012b40b9b0defe7867eaabe35998fa037312da7b903be27c23283f1b795ca WatchSource:0}: Error finding container 810012b40b9b0defe7867eaabe35998fa037312da7b903be27c23283f1b795ca: Status 404 returned error can't find the container with id 810012b40b9b0defe7867eaabe35998fa037312da7b903be27c23283f1b795ca Dec 09 10:14:58 crc kubenswrapper[4824]: I1209 10:14:58.840617 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"384b8d7f-e60a-4421-855d-6d397385f7a8","Type":"ContainerStarted","Data":"810012b40b9b0defe7867eaabe35998fa037312da7b903be27c23283f1b795ca"} Dec 09 10:14:58 crc kubenswrapper[4824]: I1209 10:14:58.874528 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"889dfa64-048d-4d55-a734-5b179dfcc1f5","Type":"ContainerStarted","Data":"12c7d4f6b5c0ed53bdebb01145b627b1460af30b6bbad2f9bd5fc324da920a1f"} Dec 09 10:14:58 crc kubenswrapper[4824]: I1209 10:14:58.875760 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"889dfa64-048d-4d55-a734-5b179dfcc1f5","Type":"ContainerStarted","Data":"bfc705c8e1c2213ec386358afc1fde6e001bd5c5d5abeaaec73e13053dc09915"} Dec 09 10:14:58 crc kubenswrapper[4824]: I1209 10:14:58.930902 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.930881918 podStartE2EDuration="2.930881918s" podCreationTimestamp="2025-12-09 10:14:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:14:58.926930883 +0000 UTC m=+1655.261435550" watchObservedRunningTime="2025-12-09 10:14:58.930881918 +0000 UTC m=+1655.265386585" Dec 09 10:14:59 crc kubenswrapper[4824]: I1209 10:14:59.232053 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 09 10:14:59 crc kubenswrapper[4824]: I1209 10:14:59.233857 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 09 10:14:59 crc kubenswrapper[4824]: I1209 10:14:59.889493 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"384b8d7f-e60a-4421-855d-6d397385f7a8","Type":"ContainerStarted","Data":"83e42cbfb500fa046b54d883713a3e0fb3e759e84101124723e2810a0f44ecc7"} Dec 09 10:15:00 crc kubenswrapper[4824]: I1209 10:15:00.181869 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421255-4ll88"] Dec 09 10:15:00 crc kubenswrapper[4824]: I1209 10:15:00.188245 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421255-4ll88" Dec 09 10:15:00 crc kubenswrapper[4824]: I1209 10:15:00.193140 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421255-4ll88"] Dec 09 10:15:00 crc kubenswrapper[4824]: I1209 10:15:00.194912 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 10:15:00 crc kubenswrapper[4824]: I1209 10:15:00.195361 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 10:15:00 crc kubenswrapper[4824]: I1209 10:15:00.347976 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zstmd\" (UniqueName: \"kubernetes.io/projected/d1598e68-56dc-4d91-8e6f-0690ac3c285d-kube-api-access-zstmd\") pod \"collect-profiles-29421255-4ll88\" (UID: \"d1598e68-56dc-4d91-8e6f-0690ac3c285d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421255-4ll88" Dec 09 10:15:00 crc kubenswrapper[4824]: I1209 10:15:00.348546 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d1598e68-56dc-4d91-8e6f-0690ac3c285d-secret-volume\") pod \"collect-profiles-29421255-4ll88\" (UID: \"d1598e68-56dc-4d91-8e6f-0690ac3c285d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421255-4ll88" Dec 09 10:15:00 crc kubenswrapper[4824]: I1209 10:15:00.348593 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1598e68-56dc-4d91-8e6f-0690ac3c285d-config-volume\") pod \"collect-profiles-29421255-4ll88\" (UID: \"d1598e68-56dc-4d91-8e6f-0690ac3c285d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421255-4ll88" Dec 09 10:15:00 crc kubenswrapper[4824]: I1209 10:15:00.371280 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:15:00 crc kubenswrapper[4824]: I1209 10:15:00.371697 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" containerName="ceilometer-central-agent" containerID="cri-o://7933eaa4a009a5ab39df47008d7965641e517ebececf33bf9eb8ecd85f7cdc0f" gracePeriod=30 Dec 09 10:15:00 crc kubenswrapper[4824]: I1209 10:15:00.372136 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" containerName="proxy-httpd" containerID="cri-o://0a9b3999d41865e7b6906497b24f006acab61039d97dccd40fe6172382468d54" gracePeriod=30 Dec 09 10:15:00 crc kubenswrapper[4824]: I1209 10:15:00.372401 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" containerName="ceilometer-notification-agent" containerID="cri-o://6f481b7674fd6b82084202baf7a20c25a6ca710f3a119487376ed5729d517bd2" gracePeriod=30 Dec 09 10:15:00 crc kubenswrapper[4824]: I1209 10:15:00.374332 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" containerName="sg-core" containerID="cri-o://f0052a0f9a94b77f605971d2f4d65e66b5cb33beee7c8b898c44b20df19c263c" gracePeriod=30 Dec 09 10:15:00 crc kubenswrapper[4824]: I1209 10:15:00.451190 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d1598e68-56dc-4d91-8e6f-0690ac3c285d-secret-volume\") pod \"collect-profiles-29421255-4ll88\" (UID: \"d1598e68-56dc-4d91-8e6f-0690ac3c285d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421255-4ll88" Dec 09 10:15:00 crc kubenswrapper[4824]: I1209 10:15:00.451278 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1598e68-56dc-4d91-8e6f-0690ac3c285d-config-volume\") pod \"collect-profiles-29421255-4ll88\" (UID: \"d1598e68-56dc-4d91-8e6f-0690ac3c285d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421255-4ll88" Dec 09 10:15:00 crc kubenswrapper[4824]: I1209 10:15:00.451370 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zstmd\" (UniqueName: \"kubernetes.io/projected/d1598e68-56dc-4d91-8e6f-0690ac3c285d-kube-api-access-zstmd\") pod \"collect-profiles-29421255-4ll88\" (UID: \"d1598e68-56dc-4d91-8e6f-0690ac3c285d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421255-4ll88" Dec 09 10:15:00 crc kubenswrapper[4824]: I1209 10:15:00.452529 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1598e68-56dc-4d91-8e6f-0690ac3c285d-config-volume\") pod \"collect-profiles-29421255-4ll88\" (UID: \"d1598e68-56dc-4d91-8e6f-0690ac3c285d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421255-4ll88" Dec 09 10:15:00 crc kubenswrapper[4824]: I1209 10:15:00.460022 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d1598e68-56dc-4d91-8e6f-0690ac3c285d-secret-volume\") pod \"collect-profiles-29421255-4ll88\" (UID: \"d1598e68-56dc-4d91-8e6f-0690ac3c285d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421255-4ll88" Dec 09 10:15:00 crc kubenswrapper[4824]: I1209 10:15:00.480673 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zstmd\" (UniqueName: \"kubernetes.io/projected/d1598e68-56dc-4d91-8e6f-0690ac3c285d-kube-api-access-zstmd\") pod \"collect-profiles-29421255-4ll88\" (UID: \"d1598e68-56dc-4d91-8e6f-0690ac3c285d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421255-4ll88" Dec 09 10:15:00 crc kubenswrapper[4824]: I1209 10:15:00.575625 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421255-4ll88" Dec 09 10:15:00 crc kubenswrapper[4824]: I1209 10:15:00.935023 4824 generic.go:334] "Generic (PLEG): container finished" podID="2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" containerID="0a9b3999d41865e7b6906497b24f006acab61039d97dccd40fe6172382468d54" exitCode=0 Dec 09 10:15:00 crc kubenswrapper[4824]: I1209 10:15:00.935489 4824 generic.go:334] "Generic (PLEG): container finished" podID="2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" containerID="f0052a0f9a94b77f605971d2f4d65e66b5cb33beee7c8b898c44b20df19c263c" exitCode=2 Dec 09 10:15:00 crc kubenswrapper[4824]: I1209 10:15:00.935732 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc","Type":"ContainerDied","Data":"0a9b3999d41865e7b6906497b24f006acab61039d97dccd40fe6172382468d54"} Dec 09 10:15:00 crc kubenswrapper[4824]: I1209 10:15:00.935804 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc","Type":"ContainerDied","Data":"f0052a0f9a94b77f605971d2f4d65e66b5cb33beee7c8b898c44b20df19c263c"} Dec 09 10:15:01 crc kubenswrapper[4824]: I1209 10:15:01.238389 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421255-4ll88"] Dec 09 10:15:01 crc kubenswrapper[4824]: I1209 10:15:01.309812 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 09 10:15:01 crc kubenswrapper[4824]: W1209 10:15:01.645746 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1598e68_56dc_4d91_8e6f_0690ac3c285d.slice/crio-65c29e67afd6ed76fff0957792895f316ac70034090f982546afffbb237d3745 WatchSource:0}: Error finding container 65c29e67afd6ed76fff0957792895f316ac70034090f982546afffbb237d3745: Status 404 returned error can't find the container with id 65c29e67afd6ed76fff0957792895f316ac70034090f982546afffbb237d3745 Dec 09 10:15:01 crc kubenswrapper[4824]: I1209 10:15:01.955493 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421255-4ll88" event={"ID":"d1598e68-56dc-4d91-8e6f-0690ac3c285d","Type":"ContainerStarted","Data":"65c29e67afd6ed76fff0957792895f316ac70034090f982546afffbb237d3745"} Dec 09 10:15:01 crc kubenswrapper[4824]: I1209 10:15:01.984221 4824 generic.go:334] "Generic (PLEG): container finished" podID="2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" containerID="7933eaa4a009a5ab39df47008d7965641e517ebececf33bf9eb8ecd85f7cdc0f" exitCode=0 Dec 09 10:15:01 crc kubenswrapper[4824]: I1209 10:15:01.984268 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc","Type":"ContainerDied","Data":"7933eaa4a009a5ab39df47008d7965641e517ebececf33bf9eb8ecd85f7cdc0f"} Dec 09 10:15:02 crc kubenswrapper[4824]: I1209 10:15:02.329016 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 09 10:15:02 crc kubenswrapper[4824]: I1209 10:15:02.860677 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:15:02 crc kubenswrapper[4824]: I1209 10:15:02.860738 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:15:03 crc kubenswrapper[4824]: I1209 10:15:03.003190 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"384b8d7f-e60a-4421-855d-6d397385f7a8","Type":"ContainerStarted","Data":"7a681790e04916ca576ca683ac87cc0c9fdee0b5b4671a1988993676391802ec"} Dec 09 10:15:03 crc kubenswrapper[4824]: I1209 10:15:03.006990 4824 generic.go:334] "Generic (PLEG): container finished" podID="d1598e68-56dc-4d91-8e6f-0690ac3c285d" containerID="487c74959ee5ed790fab7268ff15b47a012b2c30eb1e5683d571890583b9b63c" exitCode=0 Dec 09 10:15:03 crc kubenswrapper[4824]: I1209 10:15:03.007032 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421255-4ll88" event={"ID":"d1598e68-56dc-4d91-8e6f-0690ac3c285d","Type":"ContainerDied","Data":"487c74959ee5ed790fab7268ff15b47a012b2c30eb1e5683d571890583b9b63c"} Dec 09 10:15:04 crc kubenswrapper[4824]: I1209 10:15:04.049057 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"384b8d7f-e60a-4421-855d-6d397385f7a8","Type":"ContainerStarted","Data":"150ba3aafdd9b74a31610254498fb25cc5221404d539d30379abeb1ea5efa048"} Dec 09 10:15:04 crc kubenswrapper[4824]: I1209 10:15:04.239592 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 09 10:15:04 crc kubenswrapper[4824]: I1209 10:15:04.239637 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 09 10:15:04 crc kubenswrapper[4824]: I1209 10:15:04.561719 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421255-4ll88" Dec 09 10:15:04 crc kubenswrapper[4824]: I1209 10:15:04.696980 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1598e68-56dc-4d91-8e6f-0690ac3c285d-config-volume\") pod \"d1598e68-56dc-4d91-8e6f-0690ac3c285d\" (UID: \"d1598e68-56dc-4d91-8e6f-0690ac3c285d\") " Dec 09 10:15:04 crc kubenswrapper[4824]: I1209 10:15:04.697084 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zstmd\" (UniqueName: \"kubernetes.io/projected/d1598e68-56dc-4d91-8e6f-0690ac3c285d-kube-api-access-zstmd\") pod \"d1598e68-56dc-4d91-8e6f-0690ac3c285d\" (UID: \"d1598e68-56dc-4d91-8e6f-0690ac3c285d\") " Dec 09 10:15:04 crc kubenswrapper[4824]: I1209 10:15:04.697452 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d1598e68-56dc-4d91-8e6f-0690ac3c285d-secret-volume\") pod \"d1598e68-56dc-4d91-8e6f-0690ac3c285d\" (UID: \"d1598e68-56dc-4d91-8e6f-0690ac3c285d\") " Dec 09 10:15:04 crc kubenswrapper[4824]: I1209 10:15:04.698505 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1598e68-56dc-4d91-8e6f-0690ac3c285d-config-volume" (OuterVolumeSpecName: "config-volume") pod "d1598e68-56dc-4d91-8e6f-0690ac3c285d" (UID: "d1598e68-56dc-4d91-8e6f-0690ac3c285d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:15:04 crc kubenswrapper[4824]: I1209 10:15:04.756107 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1598e68-56dc-4d91-8e6f-0690ac3c285d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d1598e68-56dc-4d91-8e6f-0690ac3c285d" (UID: "d1598e68-56dc-4d91-8e6f-0690ac3c285d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:04 crc kubenswrapper[4824]: I1209 10:15:04.756235 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1598e68-56dc-4d91-8e6f-0690ac3c285d-kube-api-access-zstmd" (OuterVolumeSpecName: "kube-api-access-zstmd") pod "d1598e68-56dc-4d91-8e6f-0690ac3c285d" (UID: "d1598e68-56dc-4d91-8e6f-0690ac3c285d"). InnerVolumeSpecName "kube-api-access-zstmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:15:04 crc kubenswrapper[4824]: I1209 10:15:04.801840 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1598e68-56dc-4d91-8e6f-0690ac3c285d-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:04 crc kubenswrapper[4824]: I1209 10:15:04.802094 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zstmd\" (UniqueName: \"kubernetes.io/projected/d1598e68-56dc-4d91-8e6f-0690ac3c285d-kube-api-access-zstmd\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:04 crc kubenswrapper[4824]: I1209 10:15:04.802178 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d1598e68-56dc-4d91-8e6f-0690ac3c285d-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:05 crc kubenswrapper[4824]: I1209 10:15:05.064545 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421255-4ll88" event={"ID":"d1598e68-56dc-4d91-8e6f-0690ac3c285d","Type":"ContainerDied","Data":"65c29e67afd6ed76fff0957792895f316ac70034090f982546afffbb237d3745"} Dec 09 10:15:05 crc kubenswrapper[4824]: I1209 10:15:05.064599 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65c29e67afd6ed76fff0957792895f316ac70034090f982546afffbb237d3745" Dec 09 10:15:05 crc kubenswrapper[4824]: I1209 10:15:05.064668 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421255-4ll88" Dec 09 10:15:05 crc kubenswrapper[4824]: I1209 10:15:05.262001 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.247:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 10:15:05 crc kubenswrapper[4824]: I1209 10:15:05.262077 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.247:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 10:15:05 crc kubenswrapper[4824]: I1209 10:15:05.698800 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 09 10:15:05 crc kubenswrapper[4824]: I1209 10:15:05.698870 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.110759 4824 generic.go:334] "Generic (PLEG): container finished" podID="2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" containerID="6f481b7674fd6b82084202baf7a20c25a6ca710f3a119487376ed5729d517bd2" exitCode=0 Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.111144 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc","Type":"ContainerDied","Data":"6f481b7674fd6b82084202baf7a20c25a6ca710f3a119487376ed5729d517bd2"} Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.339494 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.496246 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" (UID: "2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.496327 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-run-httpd\") pod \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.496420 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwqh5\" (UniqueName: \"kubernetes.io/projected/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-kube-api-access-mwqh5\") pod \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.496479 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-sg-core-conf-yaml\") pod \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.497006 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" (UID: "2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.497107 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-log-httpd\") pod \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.497199 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-combined-ca-bundle\") pod \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.497226 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-scripts\") pod \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.497491 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-config-data\") pod \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\" (UID: \"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc\") " Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.498513 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.498539 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.508917 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-kube-api-access-mwqh5" (OuterVolumeSpecName: "kube-api-access-mwqh5") pod "2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" (UID: "2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc"). InnerVolumeSpecName "kube-api-access-mwqh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.512199 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-scripts" (OuterVolumeSpecName: "scripts") pod "2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" (UID: "2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.548979 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" (UID: "2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.602505 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwqh5\" (UniqueName: \"kubernetes.io/projected/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-kube-api-access-mwqh5\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.602575 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.602592 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.618865 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" (UID: "2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.653201 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-config-data" (OuterVolumeSpecName: "config-data") pod "2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" (UID: "2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.705762 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.705808 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.785064 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="74053563-b115-4fb2-9433-16e81288df8d" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.248:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 10:15:06 crc kubenswrapper[4824]: I1209 10:15:06.785360 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="74053563-b115-4fb2-9433-16e81288df8d" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.248:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.126770 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.126834 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc","Type":"ContainerDied","Data":"fc355b2d4822ce3418c6d64d86a4f386510668e8f4fbf6826a4789f89c531b97"} Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.127270 4824 scope.go:117] "RemoveContainer" containerID="0a9b3999d41865e7b6906497b24f006acab61039d97dccd40fe6172382468d54" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.138758 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"384b8d7f-e60a-4421-855d-6d397385f7a8","Type":"ContainerStarted","Data":"b96320be5c824b5fba1b470169895191d2b61ec0a5840ffe6704738b2b7bd816"} Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.138995 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="384b8d7f-e60a-4421-855d-6d397385f7a8" containerName="aodh-api" containerID="cri-o://83e42cbfb500fa046b54d883713a3e0fb3e759e84101124723e2810a0f44ecc7" gracePeriod=30 Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.139595 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="384b8d7f-e60a-4421-855d-6d397385f7a8" containerName="aodh-evaluator" containerID="cri-o://7a681790e04916ca576ca683ac87cc0c9fdee0b5b4671a1988993676391802ec" gracePeriod=30 Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.139659 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="384b8d7f-e60a-4421-855d-6d397385f7a8" containerName="aodh-notifier" containerID="cri-o://150ba3aafdd9b74a31610254498fb25cc5221404d539d30379abeb1ea5efa048" gracePeriod=30 Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.139874 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="384b8d7f-e60a-4421-855d-6d397385f7a8" containerName="aodh-listener" containerID="cri-o://b96320be5c824b5fba1b470169895191d2b61ec0a5840ffe6704738b2b7bd816" gracePeriod=30 Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.170998 4824 scope.go:117] "RemoveContainer" containerID="f0052a0f9a94b77f605971d2f4d65e66b5cb33beee7c8b898c44b20df19c263c" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.232333 4824 scope.go:117] "RemoveContainer" containerID="6f481b7674fd6b82084202baf7a20c25a6ca710f3a119487376ed5729d517bd2" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.266058 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=3.194181154 podStartE2EDuration="10.266034512s" podCreationTimestamp="2025-12-09 10:14:57 +0000 UTC" firstStartedPulling="2025-12-09 10:14:58.699965222 +0000 UTC m=+1655.034469889" lastFinishedPulling="2025-12-09 10:15:05.77181857 +0000 UTC m=+1662.106323247" observedRunningTime="2025-12-09 10:15:07.200689508 +0000 UTC m=+1663.535194185" watchObservedRunningTime="2025-12-09 10:15:07.266034512 +0000 UTC m=+1663.600539169" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.295858 4824 scope.go:117] "RemoveContainer" containerID="7933eaa4a009a5ab39df47008d7965641e517ebececf33bf9eb8ecd85f7cdc0f" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.329535 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.347196 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.360656 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.379916 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.384907 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:15:07 crc kubenswrapper[4824]: E1209 10:15:07.385698 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" containerName="ceilometer-central-agent" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.385729 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" containerName="ceilometer-central-agent" Dec 09 10:15:07 crc kubenswrapper[4824]: E1209 10:15:07.385763 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" containerName="sg-core" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.385773 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" containerName="sg-core" Dec 09 10:15:07 crc kubenswrapper[4824]: E1209 10:15:07.385818 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" containerName="ceilometer-notification-agent" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.385828 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" containerName="ceilometer-notification-agent" Dec 09 10:15:07 crc kubenswrapper[4824]: E1209 10:15:07.385846 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1598e68-56dc-4d91-8e6f-0690ac3c285d" containerName="collect-profiles" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.385854 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1598e68-56dc-4d91-8e6f-0690ac3c285d" containerName="collect-profiles" Dec 09 10:15:07 crc kubenswrapper[4824]: E1209 10:15:07.385874 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" containerName="proxy-httpd" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.385883 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" containerName="proxy-httpd" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.386225 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" containerName="proxy-httpd" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.386259 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" containerName="sg-core" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.386282 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" containerName="ceilometer-notification-agent" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.386296 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1598e68-56dc-4d91-8e6f-0690ac3c285d" containerName="collect-profiles" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.386310 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" containerName="ceilometer-central-agent" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.390658 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.394279 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.394472 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.404392 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.454109 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/391e341b-3617-4dbf-b154-b638ba7f712a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " pod="openstack/ceilometer-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.454198 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/391e341b-3617-4dbf-b154-b638ba7f712a-config-data\") pod \"ceilometer-0\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " pod="openstack/ceilometer-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.454340 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/391e341b-3617-4dbf-b154-b638ba7f712a-scripts\") pod \"ceilometer-0\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " pod="openstack/ceilometer-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.454363 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/391e341b-3617-4dbf-b154-b638ba7f712a-run-httpd\") pod \"ceilometer-0\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " pod="openstack/ceilometer-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.454398 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/391e341b-3617-4dbf-b154-b638ba7f712a-log-httpd\") pod \"ceilometer-0\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " pod="openstack/ceilometer-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.454449 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xtgm\" (UniqueName: \"kubernetes.io/projected/391e341b-3617-4dbf-b154-b638ba7f712a-kube-api-access-5xtgm\") pod \"ceilometer-0\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " pod="openstack/ceilometer-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.454869 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/391e341b-3617-4dbf-b154-b638ba7f712a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " pod="openstack/ceilometer-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.556971 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/391e341b-3617-4dbf-b154-b638ba7f712a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " pod="openstack/ceilometer-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.557172 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/391e341b-3617-4dbf-b154-b638ba7f712a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " pod="openstack/ceilometer-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.557207 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/391e341b-3617-4dbf-b154-b638ba7f712a-config-data\") pod \"ceilometer-0\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " pod="openstack/ceilometer-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.557309 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/391e341b-3617-4dbf-b154-b638ba7f712a-scripts\") pod \"ceilometer-0\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " pod="openstack/ceilometer-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.557329 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/391e341b-3617-4dbf-b154-b638ba7f712a-run-httpd\") pod \"ceilometer-0\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " pod="openstack/ceilometer-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.557387 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/391e341b-3617-4dbf-b154-b638ba7f712a-log-httpd\") pod \"ceilometer-0\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " pod="openstack/ceilometer-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.557418 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xtgm\" (UniqueName: \"kubernetes.io/projected/391e341b-3617-4dbf-b154-b638ba7f712a-kube-api-access-5xtgm\") pod \"ceilometer-0\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " pod="openstack/ceilometer-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.558912 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/391e341b-3617-4dbf-b154-b638ba7f712a-run-httpd\") pod \"ceilometer-0\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " pod="openstack/ceilometer-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.562035 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/391e341b-3617-4dbf-b154-b638ba7f712a-log-httpd\") pod \"ceilometer-0\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " pod="openstack/ceilometer-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.564284 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/391e341b-3617-4dbf-b154-b638ba7f712a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " pod="openstack/ceilometer-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.564475 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/391e341b-3617-4dbf-b154-b638ba7f712a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " pod="openstack/ceilometer-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.564579 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/391e341b-3617-4dbf-b154-b638ba7f712a-scripts\") pod \"ceilometer-0\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " pod="openstack/ceilometer-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.573468 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/391e341b-3617-4dbf-b154-b638ba7f712a-config-data\") pod \"ceilometer-0\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " pod="openstack/ceilometer-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.580572 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xtgm\" (UniqueName: \"kubernetes.io/projected/391e341b-3617-4dbf-b154-b638ba7f712a-kube-api-access-5xtgm\") pod \"ceilometer-0\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " pod="openstack/ceilometer-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.736397 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:15:07 crc kubenswrapper[4824]: I1209 10:15:07.925288 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc" path="/var/lib/kubelet/pods/2e3fdb2d-11f2-4f33-bd13-447d0cee7cfc/volumes" Dec 09 10:15:08 crc kubenswrapper[4824]: I1209 10:15:08.172609 4824 generic.go:334] "Generic (PLEG): container finished" podID="384b8d7f-e60a-4421-855d-6d397385f7a8" containerID="150ba3aafdd9b74a31610254498fb25cc5221404d539d30379abeb1ea5efa048" exitCode=0 Dec 09 10:15:08 crc kubenswrapper[4824]: I1209 10:15:08.172931 4824 generic.go:334] "Generic (PLEG): container finished" podID="384b8d7f-e60a-4421-855d-6d397385f7a8" containerID="7a681790e04916ca576ca683ac87cc0c9fdee0b5b4671a1988993676391802ec" exitCode=0 Dec 09 10:15:08 crc kubenswrapper[4824]: I1209 10:15:08.172945 4824 generic.go:334] "Generic (PLEG): container finished" podID="384b8d7f-e60a-4421-855d-6d397385f7a8" containerID="83e42cbfb500fa046b54d883713a3e0fb3e759e84101124723e2810a0f44ecc7" exitCode=0 Dec 09 10:15:08 crc kubenswrapper[4824]: I1209 10:15:08.173001 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"384b8d7f-e60a-4421-855d-6d397385f7a8","Type":"ContainerDied","Data":"150ba3aafdd9b74a31610254498fb25cc5221404d539d30379abeb1ea5efa048"} Dec 09 10:15:08 crc kubenswrapper[4824]: I1209 10:15:08.173032 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"384b8d7f-e60a-4421-855d-6d397385f7a8","Type":"ContainerDied","Data":"7a681790e04916ca576ca683ac87cc0c9fdee0b5b4671a1988993676391802ec"} Dec 09 10:15:08 crc kubenswrapper[4824]: I1209 10:15:08.173043 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"384b8d7f-e60a-4421-855d-6d397385f7a8","Type":"ContainerDied","Data":"83e42cbfb500fa046b54d883713a3e0fb3e759e84101124723e2810a0f44ecc7"} Dec 09 10:15:08 crc kubenswrapper[4824]: I1209 10:15:08.231029 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 09 10:15:08 crc kubenswrapper[4824]: I1209 10:15:08.419971 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:15:08 crc kubenswrapper[4824]: E1209 10:15:08.606734 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb97d1be8_3f68_488f_a879_ae2ab5b44cfb.slice/crio-conmon-671b4caf32e6e41880e92594a0ad626b69af708fa7f1ccd387e7fdd7f87aaa0f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb97d1be8_3f68_488f_a879_ae2ab5b44cfb.slice/crio-671b4caf32e6e41880e92594a0ad626b69af708fa7f1ccd387e7fdd7f87aaa0f.scope\": RecentStats: unable to find data in memory cache]" Dec 09 10:15:09 crc kubenswrapper[4824]: I1209 10:15:09.192547 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"391e341b-3617-4dbf-b154-b638ba7f712a","Type":"ContainerStarted","Data":"2a2ca948433846eaacf439c1f673373ed07bc59622bdfa08584a196785673e59"} Dec 09 10:15:09 crc kubenswrapper[4824]: I1209 10:15:09.198104 4824 generic.go:334] "Generic (PLEG): container finished" podID="b97d1be8-3f68-488f-a879-ae2ab5b44cfb" containerID="671b4caf32e6e41880e92594a0ad626b69af708fa7f1ccd387e7fdd7f87aaa0f" exitCode=137 Dec 09 10:15:09 crc kubenswrapper[4824]: I1209 10:15:09.199569 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b97d1be8-3f68-488f-a879-ae2ab5b44cfb","Type":"ContainerDied","Data":"671b4caf32e6e41880e92594a0ad626b69af708fa7f1ccd387e7fdd7f87aaa0f"} Dec 09 10:15:09 crc kubenswrapper[4824]: I1209 10:15:09.199606 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b97d1be8-3f68-488f-a879-ae2ab5b44cfb","Type":"ContainerDied","Data":"ff86603f59e1e5c9da9104481b88a3201327044676b4755fc48536340584549a"} Dec 09 10:15:09 crc kubenswrapper[4824]: I1209 10:15:09.199619 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff86603f59e1e5c9da9104481b88a3201327044676b4755fc48536340584549a" Dec 09 10:15:09 crc kubenswrapper[4824]: I1209 10:15:09.334213 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:15:09 crc kubenswrapper[4824]: I1209 10:15:09.517039 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b97d1be8-3f68-488f-a879-ae2ab5b44cfb-combined-ca-bundle\") pod \"b97d1be8-3f68-488f-a879-ae2ab5b44cfb\" (UID: \"b97d1be8-3f68-488f-a879-ae2ab5b44cfb\") " Dec 09 10:15:09 crc kubenswrapper[4824]: I1209 10:15:09.517106 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b97d1be8-3f68-488f-a879-ae2ab5b44cfb-config-data\") pod \"b97d1be8-3f68-488f-a879-ae2ab5b44cfb\" (UID: \"b97d1be8-3f68-488f-a879-ae2ab5b44cfb\") " Dec 09 10:15:09 crc kubenswrapper[4824]: I1209 10:15:09.519362 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2kd5\" (UniqueName: \"kubernetes.io/projected/b97d1be8-3f68-488f-a879-ae2ab5b44cfb-kube-api-access-z2kd5\") pod \"b97d1be8-3f68-488f-a879-ae2ab5b44cfb\" (UID: \"b97d1be8-3f68-488f-a879-ae2ab5b44cfb\") " Dec 09 10:15:09 crc kubenswrapper[4824]: I1209 10:15:09.533818 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b97d1be8-3f68-488f-a879-ae2ab5b44cfb-kube-api-access-z2kd5" (OuterVolumeSpecName: "kube-api-access-z2kd5") pod "b97d1be8-3f68-488f-a879-ae2ab5b44cfb" (UID: "b97d1be8-3f68-488f-a879-ae2ab5b44cfb"). InnerVolumeSpecName "kube-api-access-z2kd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:15:09 crc kubenswrapper[4824]: I1209 10:15:09.560498 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b97d1be8-3f68-488f-a879-ae2ab5b44cfb-config-data" (OuterVolumeSpecName: "config-data") pod "b97d1be8-3f68-488f-a879-ae2ab5b44cfb" (UID: "b97d1be8-3f68-488f-a879-ae2ab5b44cfb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:09 crc kubenswrapper[4824]: I1209 10:15:09.561137 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b97d1be8-3f68-488f-a879-ae2ab5b44cfb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b97d1be8-3f68-488f-a879-ae2ab5b44cfb" (UID: "b97d1be8-3f68-488f-a879-ae2ab5b44cfb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:09 crc kubenswrapper[4824]: I1209 10:15:09.623929 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b97d1be8-3f68-488f-a879-ae2ab5b44cfb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:09 crc kubenswrapper[4824]: I1209 10:15:09.623980 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b97d1be8-3f68-488f-a879-ae2ab5b44cfb-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:09 crc kubenswrapper[4824]: I1209 10:15:09.623992 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2kd5\" (UniqueName: \"kubernetes.io/projected/b97d1be8-3f68-488f-a879-ae2ab5b44cfb-kube-api-access-z2kd5\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.214382 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.215284 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="94d75a5f-1730-4993-9948-c42c98719163" containerName="kube-state-metrics" containerID="cri-o://70f793c8069bc8d2b0fd5823635cb4392135e5e442704367b9f8a994600336f5" gracePeriod=30 Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.230593 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.231544 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"391e341b-3617-4dbf-b154-b638ba7f712a","Type":"ContainerStarted","Data":"e41b24ea7f30d94712a8e4178494349c8616a0097dd1661305970dc477011c67"} Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.289429 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.304522 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.330911 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 10:15:10 crc kubenswrapper[4824]: E1209 10:15:10.332048 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b97d1be8-3f68-488f-a879-ae2ab5b44cfb" containerName="nova-cell1-novncproxy-novncproxy" Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.332079 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b97d1be8-3f68-488f-a879-ae2ab5b44cfb" containerName="nova-cell1-novncproxy-novncproxy" Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.332376 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b97d1be8-3f68-488f-a879-ae2ab5b44cfb" containerName="nova-cell1-novncproxy-novncproxy" Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.333908 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.340255 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.340550 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.340682 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.351903 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.448705 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6208994-9aba-4007-ba54-b8718e789f4b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a6208994-9aba-4007-ba54-b8718e789f4b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.449515 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6208994-9aba-4007-ba54-b8718e789f4b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a6208994-9aba-4007-ba54-b8718e789f4b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.450149 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6208994-9aba-4007-ba54-b8718e789f4b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a6208994-9aba-4007-ba54-b8718e789f4b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.452153 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6208994-9aba-4007-ba54-b8718e789f4b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a6208994-9aba-4007-ba54-b8718e789f4b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.452392 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lpcp\" (UniqueName: \"kubernetes.io/projected/a6208994-9aba-4007-ba54-b8718e789f4b-kube-api-access-8lpcp\") pod \"nova-cell1-novncproxy-0\" (UID: \"a6208994-9aba-4007-ba54-b8718e789f4b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.537953 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.538366 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mysqld-exporter-0" podUID="8d47a992-6428-4729-8f4d-04af792e688e" containerName="mysqld-exporter" containerID="cri-o://2d99e9618d5d9984148ea2573701fb0fd09b8bb980843ed57ac98b1dd5b4c3e7" gracePeriod=30 Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.560157 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6208994-9aba-4007-ba54-b8718e789f4b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a6208994-9aba-4007-ba54-b8718e789f4b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.560268 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6208994-9aba-4007-ba54-b8718e789f4b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a6208994-9aba-4007-ba54-b8718e789f4b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.560303 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lpcp\" (UniqueName: \"kubernetes.io/projected/a6208994-9aba-4007-ba54-b8718e789f4b-kube-api-access-8lpcp\") pod \"nova-cell1-novncproxy-0\" (UID: \"a6208994-9aba-4007-ba54-b8718e789f4b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.560491 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6208994-9aba-4007-ba54-b8718e789f4b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a6208994-9aba-4007-ba54-b8718e789f4b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.563359 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6208994-9aba-4007-ba54-b8718e789f4b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a6208994-9aba-4007-ba54-b8718e789f4b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.571899 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6208994-9aba-4007-ba54-b8718e789f4b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a6208994-9aba-4007-ba54-b8718e789f4b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.572819 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6208994-9aba-4007-ba54-b8718e789f4b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a6208994-9aba-4007-ba54-b8718e789f4b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.575845 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6208994-9aba-4007-ba54-b8718e789f4b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a6208994-9aba-4007-ba54-b8718e789f4b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.576305 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6208994-9aba-4007-ba54-b8718e789f4b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a6208994-9aba-4007-ba54-b8718e789f4b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.589897 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lpcp\" (UniqueName: \"kubernetes.io/projected/a6208994-9aba-4007-ba54-b8718e789f4b-kube-api-access-8lpcp\") pod \"nova-cell1-novncproxy-0\" (UID: \"a6208994-9aba-4007-ba54-b8718e789f4b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:15:10 crc kubenswrapper[4824]: I1209 10:15:10.743030 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.183769 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.260950 4824 generic.go:334] "Generic (PLEG): container finished" podID="94d75a5f-1730-4993-9948-c42c98719163" containerID="70f793c8069bc8d2b0fd5823635cb4392135e5e442704367b9f8a994600336f5" exitCode=2 Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.261096 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"94d75a5f-1730-4993-9948-c42c98719163","Type":"ContainerDied","Data":"70f793c8069bc8d2b0fd5823635cb4392135e5e442704367b9f8a994600336f5"} Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.261148 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"94d75a5f-1730-4993-9948-c42c98719163","Type":"ContainerDied","Data":"f61ec0a06ab6a4d4661310bad885a53a690ad0a6a4b424db05c42e6e27e1a55b"} Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.261168 4824 scope.go:117] "RemoveContainer" containerID="70f793c8069bc8d2b0fd5823635cb4392135e5e442704367b9f8a994600336f5" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.261419 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.269937 4824 generic.go:334] "Generic (PLEG): container finished" podID="8d47a992-6428-4729-8f4d-04af792e688e" containerID="2d99e9618d5d9984148ea2573701fb0fd09b8bb980843ed57ac98b1dd5b4c3e7" exitCode=2 Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.270009 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"8d47a992-6428-4729-8f4d-04af792e688e","Type":"ContainerDied","Data":"2d99e9618d5d9984148ea2573701fb0fd09b8bb980843ed57ac98b1dd5b4c3e7"} Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.276066 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"391e341b-3617-4dbf-b154-b638ba7f712a","Type":"ContainerStarted","Data":"c67996f5021ab1519dc46ae411b48b0056f680f5e611031d5dc7e262e967a3da"} Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.306676 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsvhp\" (UniqueName: \"kubernetes.io/projected/94d75a5f-1730-4993-9948-c42c98719163-kube-api-access-hsvhp\") pod \"94d75a5f-1730-4993-9948-c42c98719163\" (UID: \"94d75a5f-1730-4993-9948-c42c98719163\") " Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.322351 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94d75a5f-1730-4993-9948-c42c98719163-kube-api-access-hsvhp" (OuterVolumeSpecName: "kube-api-access-hsvhp") pod "94d75a5f-1730-4993-9948-c42c98719163" (UID: "94d75a5f-1730-4993-9948-c42c98719163"). InnerVolumeSpecName "kube-api-access-hsvhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.334279 4824 scope.go:117] "RemoveContainer" containerID="70f793c8069bc8d2b0fd5823635cb4392135e5e442704367b9f8a994600336f5" Dec 09 10:15:11 crc kubenswrapper[4824]: E1209 10:15:11.338407 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70f793c8069bc8d2b0fd5823635cb4392135e5e442704367b9f8a994600336f5\": container with ID starting with 70f793c8069bc8d2b0fd5823635cb4392135e5e442704367b9f8a994600336f5 not found: ID does not exist" containerID="70f793c8069bc8d2b0fd5823635cb4392135e5e442704367b9f8a994600336f5" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.338452 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70f793c8069bc8d2b0fd5823635cb4392135e5e442704367b9f8a994600336f5"} err="failed to get container status \"70f793c8069bc8d2b0fd5823635cb4392135e5e442704367b9f8a994600336f5\": rpc error: code = NotFound desc = could not find container \"70f793c8069bc8d2b0fd5823635cb4392135e5e442704367b9f8a994600336f5\": container with ID starting with 70f793c8069bc8d2b0fd5823635cb4392135e5e442704367b9f8a994600336f5 not found: ID does not exist" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.415039 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsvhp\" (UniqueName: \"kubernetes.io/projected/94d75a5f-1730-4993-9948-c42c98719163-kube-api-access-hsvhp\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.441954 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.618713 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.627324 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d47a992-6428-4729-8f4d-04af792e688e-combined-ca-bundle\") pod \"8d47a992-6428-4729-8f4d-04af792e688e\" (UID: \"8d47a992-6428-4729-8f4d-04af792e688e\") " Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.627453 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d47a992-6428-4729-8f4d-04af792e688e-config-data\") pod \"8d47a992-6428-4729-8f4d-04af792e688e\" (UID: \"8d47a992-6428-4729-8f4d-04af792e688e\") " Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.627614 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltv52\" (UniqueName: \"kubernetes.io/projected/8d47a992-6428-4729-8f4d-04af792e688e-kube-api-access-ltv52\") pod \"8d47a992-6428-4729-8f4d-04af792e688e\" (UID: \"8d47a992-6428-4729-8f4d-04af792e688e\") " Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.635985 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d47a992-6428-4729-8f4d-04af792e688e-kube-api-access-ltv52" (OuterVolumeSpecName: "kube-api-access-ltv52") pod "8d47a992-6428-4729-8f4d-04af792e688e" (UID: "8d47a992-6428-4729-8f4d-04af792e688e"). InnerVolumeSpecName "kube-api-access-ltv52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.642520 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.662179 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 10:15:11 crc kubenswrapper[4824]: E1209 10:15:11.662666 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94d75a5f-1730-4993-9948-c42c98719163" containerName="kube-state-metrics" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.662683 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="94d75a5f-1730-4993-9948-c42c98719163" containerName="kube-state-metrics" Dec 09 10:15:11 crc kubenswrapper[4824]: E1209 10:15:11.662728 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d47a992-6428-4729-8f4d-04af792e688e" containerName="mysqld-exporter" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.662735 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d47a992-6428-4729-8f4d-04af792e688e" containerName="mysqld-exporter" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.663003 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="94d75a5f-1730-4993-9948-c42c98719163" containerName="kube-state-metrics" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.663023 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d47a992-6428-4729-8f4d-04af792e688e" containerName="mysqld-exporter" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.663880 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.675421 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.675629 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.679904 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.694861 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.731234 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54621281-db00-41e4-b617-032435893391-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"54621281-db00-41e4-b617-032435893391\") " pod="openstack/kube-state-metrics-0" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.731351 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nkv5\" (UniqueName: \"kubernetes.io/projected/54621281-db00-41e4-b617-032435893391-kube-api-access-5nkv5\") pod \"kube-state-metrics-0\" (UID: \"54621281-db00-41e4-b617-032435893391\") " pod="openstack/kube-state-metrics-0" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.731422 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/54621281-db00-41e4-b617-032435893391-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"54621281-db00-41e4-b617-032435893391\") " pod="openstack/kube-state-metrics-0" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.731453 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/54621281-db00-41e4-b617-032435893391-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"54621281-db00-41e4-b617-032435893391\") " pod="openstack/kube-state-metrics-0" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.731677 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltv52\" (UniqueName: \"kubernetes.io/projected/8d47a992-6428-4729-8f4d-04af792e688e-kube-api-access-ltv52\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.764830 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d47a992-6428-4729-8f4d-04af792e688e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d47a992-6428-4729-8f4d-04af792e688e" (UID: "8d47a992-6428-4729-8f4d-04af792e688e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.833510 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54621281-db00-41e4-b617-032435893391-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"54621281-db00-41e4-b617-032435893391\") " pod="openstack/kube-state-metrics-0" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.833595 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nkv5\" (UniqueName: \"kubernetes.io/projected/54621281-db00-41e4-b617-032435893391-kube-api-access-5nkv5\") pod \"kube-state-metrics-0\" (UID: \"54621281-db00-41e4-b617-032435893391\") " pod="openstack/kube-state-metrics-0" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.833651 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/54621281-db00-41e4-b617-032435893391-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"54621281-db00-41e4-b617-032435893391\") " pod="openstack/kube-state-metrics-0" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.833696 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/54621281-db00-41e4-b617-032435893391-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"54621281-db00-41e4-b617-032435893391\") " pod="openstack/kube-state-metrics-0" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.833835 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d47a992-6428-4729-8f4d-04af792e688e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.837393 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/54621281-db00-41e4-b617-032435893391-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"54621281-db00-41e4-b617-032435893391\") " pod="openstack/kube-state-metrics-0" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.843606 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54621281-db00-41e4-b617-032435893391-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"54621281-db00-41e4-b617-032435893391\") " pod="openstack/kube-state-metrics-0" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.848536 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/54621281-db00-41e4-b617-032435893391-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"54621281-db00-41e4-b617-032435893391\") " pod="openstack/kube-state-metrics-0" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.854803 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nkv5\" (UniqueName: \"kubernetes.io/projected/54621281-db00-41e4-b617-032435893391-kube-api-access-5nkv5\") pod \"kube-state-metrics-0\" (UID: \"54621281-db00-41e4-b617-032435893391\") " pod="openstack/kube-state-metrics-0" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.865936 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d47a992-6428-4729-8f4d-04af792e688e-config-data" (OuterVolumeSpecName: "config-data") pod "8d47a992-6428-4729-8f4d-04af792e688e" (UID: "8d47a992-6428-4729-8f4d-04af792e688e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.939276 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d47a992-6428-4729-8f4d-04af792e688e-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.942892 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94d75a5f-1730-4993-9948-c42c98719163" path="/var/lib/kubelet/pods/94d75a5f-1730-4993-9948-c42c98719163/volumes" Dec 09 10:15:11 crc kubenswrapper[4824]: I1209 10:15:11.943769 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b97d1be8-3f68-488f-a879-ae2ab5b44cfb" path="/var/lib/kubelet/pods/b97d1be8-3f68-488f-a879-ae2ab5b44cfb/volumes" Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.063917 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.185818 4824 scope.go:117] "RemoveContainer" containerID="7cf4f49e8f8e0800f59be8f22816ba55ee3205fc0e8f6c81552552f0e44a23f7" Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.365216 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a6208994-9aba-4007-ba54-b8718e789f4b","Type":"ContainerStarted","Data":"01bf1c55725dd87715051455c3bcff097022f5989412a123b5545dc39f05b464"} Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.381867 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.547531 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"8d47a992-6428-4729-8f4d-04af792e688e","Type":"ContainerDied","Data":"d6dc549978dfaf666b6b09ba5a2997fd6b3a47fb7a0cd29214e4fc85e0c92702"} Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.547587 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"391e341b-3617-4dbf-b154-b638ba7f712a","Type":"ContainerStarted","Data":"ea40aa4f9156ace57b9937dab91e9c78a0b2d247880d0221dbb87e4fc3685676"} Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.547615 4824 scope.go:117] "RemoveContainer" containerID="2d99e9618d5d9984148ea2573701fb0fd09b8bb980843ed57ac98b1dd5b4c3e7" Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.641886 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.658151 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.692294 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.693890 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.696506 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.698009 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-mysqld-exporter-svc" Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.745831 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.766013 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.811932 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6fkb\" (UniqueName: \"kubernetes.io/projected/674c8cb5-4f45-48a9-9b64-07ce6e0a0865-kube-api-access-r6fkb\") pod \"mysqld-exporter-0\" (UID: \"674c8cb5-4f45-48a9-9b64-07ce6e0a0865\") " pod="openstack/mysqld-exporter-0" Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.811998 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/674c8cb5-4f45-48a9-9b64-07ce6e0a0865-config-data\") pod \"mysqld-exporter-0\" (UID: \"674c8cb5-4f45-48a9-9b64-07ce6e0a0865\") " pod="openstack/mysqld-exporter-0" Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.812031 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/674c8cb5-4f45-48a9-9b64-07ce6e0a0865-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"674c8cb5-4f45-48a9-9b64-07ce6e0a0865\") " pod="openstack/mysqld-exporter-0" Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.812122 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/674c8cb5-4f45-48a9-9b64-07ce6e0a0865-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"674c8cb5-4f45-48a9-9b64-07ce6e0a0865\") " pod="openstack/mysqld-exporter-0" Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.913918 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6fkb\" (UniqueName: \"kubernetes.io/projected/674c8cb5-4f45-48a9-9b64-07ce6e0a0865-kube-api-access-r6fkb\") pod \"mysqld-exporter-0\" (UID: \"674c8cb5-4f45-48a9-9b64-07ce6e0a0865\") " pod="openstack/mysqld-exporter-0" Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.913987 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/674c8cb5-4f45-48a9-9b64-07ce6e0a0865-config-data\") pod \"mysqld-exporter-0\" (UID: \"674c8cb5-4f45-48a9-9b64-07ce6e0a0865\") " pod="openstack/mysqld-exporter-0" Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.914022 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/674c8cb5-4f45-48a9-9b64-07ce6e0a0865-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"674c8cb5-4f45-48a9-9b64-07ce6e0a0865\") " pod="openstack/mysqld-exporter-0" Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.914332 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/674c8cb5-4f45-48a9-9b64-07ce6e0a0865-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"674c8cb5-4f45-48a9-9b64-07ce6e0a0865\") " pod="openstack/mysqld-exporter-0" Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.921153 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/674c8cb5-4f45-48a9-9b64-07ce6e0a0865-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"674c8cb5-4f45-48a9-9b64-07ce6e0a0865\") " pod="openstack/mysqld-exporter-0" Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.921139 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/674c8cb5-4f45-48a9-9b64-07ce6e0a0865-config-data\") pod \"mysqld-exporter-0\" (UID: \"674c8cb5-4f45-48a9-9b64-07ce6e0a0865\") " pod="openstack/mysqld-exporter-0" Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.926903 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/674c8cb5-4f45-48a9-9b64-07ce6e0a0865-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"674c8cb5-4f45-48a9-9b64-07ce6e0a0865\") " pod="openstack/mysqld-exporter-0" Dec 09 10:15:12 crc kubenswrapper[4824]: I1209 10:15:12.950475 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6fkb\" (UniqueName: \"kubernetes.io/projected/674c8cb5-4f45-48a9-9b64-07ce6e0a0865-kube-api-access-r6fkb\") pod \"mysqld-exporter-0\" (UID: \"674c8cb5-4f45-48a9-9b64-07ce6e0a0865\") " pod="openstack/mysqld-exporter-0" Dec 09 10:15:13 crc kubenswrapper[4824]: I1209 10:15:13.114158 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 09 10:15:13 crc kubenswrapper[4824]: I1209 10:15:13.582233 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a6208994-9aba-4007-ba54-b8718e789f4b","Type":"ContainerStarted","Data":"65021d22d6ad157e78865ec816f908fdba12b0ccf4b0f796de0bd08bbdd7974d"} Dec 09 10:15:13 crc kubenswrapper[4824]: I1209 10:15:13.617037 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"54621281-db00-41e4-b617-032435893391","Type":"ContainerStarted","Data":"ac6900fe6833d0f591a1e3c5f48abd7d70fc43fcb311db3cf5ee879aa0fa111a"} Dec 09 10:15:13 crc kubenswrapper[4824]: I1209 10:15:13.622534 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.622495437 podStartE2EDuration="3.622495437s" podCreationTimestamp="2025-12-09 10:15:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:15:13.619586105 +0000 UTC m=+1669.954090772" watchObservedRunningTime="2025-12-09 10:15:13.622495437 +0000 UTC m=+1669.957000104" Dec 09 10:15:13 crc kubenswrapper[4824]: I1209 10:15:13.650285 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"391e341b-3617-4dbf-b154-b638ba7f712a","Type":"ContainerStarted","Data":"3b4fc6ee8cc316065513e16c9544a9397e005caeb346087dfa5e9025bbf2342b"} Dec 09 10:15:13 crc kubenswrapper[4824]: I1209 10:15:13.651004 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 10:15:13 crc kubenswrapper[4824]: I1209 10:15:13.691180 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.238317411 podStartE2EDuration="6.691157746s" podCreationTimestamp="2025-12-09 10:15:07 +0000 UTC" firstStartedPulling="2025-12-09 10:15:08.435510205 +0000 UTC m=+1664.770014872" lastFinishedPulling="2025-12-09 10:15:12.88835054 +0000 UTC m=+1669.222855207" observedRunningTime="2025-12-09 10:15:13.678403813 +0000 UTC m=+1670.012908490" watchObservedRunningTime="2025-12-09 10:15:13.691157746 +0000 UTC m=+1670.025662423" Dec 09 10:15:13 crc kubenswrapper[4824]: I1209 10:15:13.788379 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 09 10:15:13 crc kubenswrapper[4824]: I1209 10:15:13.961365 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d47a992-6428-4729-8f4d-04af792e688e" path="/var/lib/kubelet/pods/8d47a992-6428-4729-8f4d-04af792e688e/volumes" Dec 09 10:15:14 crc kubenswrapper[4824]: I1209 10:15:14.238385 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 09 10:15:14 crc kubenswrapper[4824]: I1209 10:15:14.239149 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 09 10:15:14 crc kubenswrapper[4824]: I1209 10:15:14.248601 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 09 10:15:14 crc kubenswrapper[4824]: I1209 10:15:14.520462 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:15:14 crc kubenswrapper[4824]: I1209 10:15:14.696346 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"54621281-db00-41e4-b617-032435893391","Type":"ContainerStarted","Data":"2bbad6105a2324d3ba9b939f829736af090521c9ba0ff0e211e1344cab369da7"} Dec 09 10:15:14 crc kubenswrapper[4824]: I1209 10:15:14.697341 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 09 10:15:14 crc kubenswrapper[4824]: I1209 10:15:14.701906 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"674c8cb5-4f45-48a9-9b64-07ce6e0a0865","Type":"ContainerStarted","Data":"a7625d5877535168123268acaa69e2bce61505ae0e462c55b84531d527ae3a90"} Dec 09 10:15:14 crc kubenswrapper[4824]: I1209 10:15:14.701960 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"674c8cb5-4f45-48a9-9b64-07ce6e0a0865","Type":"ContainerStarted","Data":"0dbf75cb4577ee576ca57edb7b9e31d56e384437f484ec862e71c8208e3c333d"} Dec 09 10:15:14 crc kubenswrapper[4824]: I1209 10:15:14.711411 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 09 10:15:14 crc kubenswrapper[4824]: I1209 10:15:14.733645 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.307071346 podStartE2EDuration="3.733622145s" podCreationTimestamp="2025-12-09 10:15:11 +0000 UTC" firstStartedPulling="2025-12-09 10:15:12.758812736 +0000 UTC m=+1669.093317413" lastFinishedPulling="2025-12-09 10:15:13.185363545 +0000 UTC m=+1669.519868212" observedRunningTime="2025-12-09 10:15:14.720254992 +0000 UTC m=+1671.054759689" watchObservedRunningTime="2025-12-09 10:15:14.733622145 +0000 UTC m=+1671.068126812" Dec 09 10:15:14 crc kubenswrapper[4824]: I1209 10:15:14.827059 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=2.33852973 podStartE2EDuration="2.827033456s" podCreationTimestamp="2025-12-09 10:15:12 +0000 UTC" firstStartedPulling="2025-12-09 10:15:13.775611334 +0000 UTC m=+1670.110116041" lastFinishedPulling="2025-12-09 10:15:14.26411511 +0000 UTC m=+1670.598619767" observedRunningTime="2025-12-09 10:15:14.775714755 +0000 UTC m=+1671.110219432" watchObservedRunningTime="2025-12-09 10:15:14.827033456 +0000 UTC m=+1671.161538123" Dec 09 10:15:15 crc kubenswrapper[4824]: I1209 10:15:15.707567 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 09 10:15:15 crc kubenswrapper[4824]: I1209 10:15:15.708200 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 09 10:15:15 crc kubenswrapper[4824]: I1209 10:15:15.710085 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="391e341b-3617-4dbf-b154-b638ba7f712a" containerName="ceilometer-central-agent" containerID="cri-o://e41b24ea7f30d94712a8e4178494349c8616a0097dd1661305970dc477011c67" gracePeriod=30 Dec 09 10:15:15 crc kubenswrapper[4824]: I1209 10:15:15.710131 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="391e341b-3617-4dbf-b154-b638ba7f712a" containerName="ceilometer-notification-agent" containerID="cri-o://c67996f5021ab1519dc46ae411b48b0056f680f5e611031d5dc7e262e967a3da" gracePeriod=30 Dec 09 10:15:15 crc kubenswrapper[4824]: I1209 10:15:15.710207 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="391e341b-3617-4dbf-b154-b638ba7f712a" containerName="sg-core" containerID="cri-o://ea40aa4f9156ace57b9937dab91e9c78a0b2d247880d0221dbb87e4fc3685676" gracePeriod=30 Dec 09 10:15:15 crc kubenswrapper[4824]: I1209 10:15:15.710095 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="391e341b-3617-4dbf-b154-b638ba7f712a" containerName="proxy-httpd" containerID="cri-o://3b4fc6ee8cc316065513e16c9544a9397e005caeb346087dfa5e9025bbf2342b" gracePeriod=30 Dec 09 10:15:15 crc kubenswrapper[4824]: I1209 10:15:15.713994 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 09 10:15:15 crc kubenswrapper[4824]: I1209 10:15:15.716305 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 09 10:15:15 crc kubenswrapper[4824]: I1209 10:15:15.744257 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:15:16 crc kubenswrapper[4824]: I1209 10:15:16.726130 4824 generic.go:334] "Generic (PLEG): container finished" podID="391e341b-3617-4dbf-b154-b638ba7f712a" containerID="3b4fc6ee8cc316065513e16c9544a9397e005caeb346087dfa5e9025bbf2342b" exitCode=0 Dec 09 10:15:16 crc kubenswrapper[4824]: I1209 10:15:16.726474 4824 generic.go:334] "Generic (PLEG): container finished" podID="391e341b-3617-4dbf-b154-b638ba7f712a" containerID="ea40aa4f9156ace57b9937dab91e9c78a0b2d247880d0221dbb87e4fc3685676" exitCode=2 Dec 09 10:15:16 crc kubenswrapper[4824]: I1209 10:15:16.726488 4824 generic.go:334] "Generic (PLEG): container finished" podID="391e341b-3617-4dbf-b154-b638ba7f712a" containerID="c67996f5021ab1519dc46ae411b48b0056f680f5e611031d5dc7e262e967a3da" exitCode=0 Dec 09 10:15:16 crc kubenswrapper[4824]: I1209 10:15:16.728211 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"391e341b-3617-4dbf-b154-b638ba7f712a","Type":"ContainerDied","Data":"3b4fc6ee8cc316065513e16c9544a9397e005caeb346087dfa5e9025bbf2342b"} Dec 09 10:15:16 crc kubenswrapper[4824]: I1209 10:15:16.728258 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 09 10:15:16 crc kubenswrapper[4824]: I1209 10:15:16.728274 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"391e341b-3617-4dbf-b154-b638ba7f712a","Type":"ContainerDied","Data":"ea40aa4f9156ace57b9937dab91e9c78a0b2d247880d0221dbb87e4fc3685676"} Dec 09 10:15:16 crc kubenswrapper[4824]: I1209 10:15:16.728287 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"391e341b-3617-4dbf-b154-b638ba7f712a","Type":"ContainerDied","Data":"c67996f5021ab1519dc46ae411b48b0056f680f5e611031d5dc7e262e967a3da"} Dec 09 10:15:16 crc kubenswrapper[4824]: I1209 10:15:16.732355 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 09 10:15:16 crc kubenswrapper[4824]: I1209 10:15:16.926176 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn"] Dec 09 10:15:16 crc kubenswrapper[4824]: I1209 10:15:16.928461 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" Dec 09 10:15:16 crc kubenswrapper[4824]: I1209 10:15:16.949881 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn"] Dec 09 10:15:16 crc kubenswrapper[4824]: I1209 10:15:16.969026 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-lhfdn\" (UID: \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" Dec 09 10:15:16 crc kubenswrapper[4824]: I1209 10:15:16.969085 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnv56\" (UniqueName: \"kubernetes.io/projected/a4e6339b-2aab-42c8-8783-5ffe08b2a026-kube-api-access-cnv56\") pod \"dnsmasq-dns-6b7bbf7cf9-lhfdn\" (UID: \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" Dec 09 10:15:16 crc kubenswrapper[4824]: I1209 10:15:16.969177 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-lhfdn\" (UID: \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" Dec 09 10:15:16 crc kubenswrapper[4824]: I1209 10:15:16.969226 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-config\") pod \"dnsmasq-dns-6b7bbf7cf9-lhfdn\" (UID: \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" Dec 09 10:15:16 crc kubenswrapper[4824]: I1209 10:15:16.969254 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-lhfdn\" (UID: \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" Dec 09 10:15:16 crc kubenswrapper[4824]: I1209 10:15:16.969300 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-lhfdn\" (UID: \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" Dec 09 10:15:17 crc kubenswrapper[4824]: I1209 10:15:17.072465 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-lhfdn\" (UID: \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" Dec 09 10:15:17 crc kubenswrapper[4824]: I1209 10:15:17.072849 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnv56\" (UniqueName: \"kubernetes.io/projected/a4e6339b-2aab-42c8-8783-5ffe08b2a026-kube-api-access-cnv56\") pod \"dnsmasq-dns-6b7bbf7cf9-lhfdn\" (UID: \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" Dec 09 10:15:17 crc kubenswrapper[4824]: I1209 10:15:17.072967 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-lhfdn\" (UID: \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" Dec 09 10:15:17 crc kubenswrapper[4824]: I1209 10:15:17.073032 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-config\") pod \"dnsmasq-dns-6b7bbf7cf9-lhfdn\" (UID: \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" Dec 09 10:15:17 crc kubenswrapper[4824]: I1209 10:15:17.073069 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-lhfdn\" (UID: \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" Dec 09 10:15:17 crc kubenswrapper[4824]: I1209 10:15:17.073125 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-lhfdn\" (UID: \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" Dec 09 10:15:17 crc kubenswrapper[4824]: I1209 10:15:17.074268 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-lhfdn\" (UID: \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" Dec 09 10:15:17 crc kubenswrapper[4824]: I1209 10:15:17.074281 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-lhfdn\" (UID: \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" Dec 09 10:15:17 crc kubenswrapper[4824]: I1209 10:15:17.074570 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-lhfdn\" (UID: \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" Dec 09 10:15:17 crc kubenswrapper[4824]: I1209 10:15:17.074798 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-lhfdn\" (UID: \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" Dec 09 10:15:17 crc kubenswrapper[4824]: I1209 10:15:17.075319 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-config\") pod \"dnsmasq-dns-6b7bbf7cf9-lhfdn\" (UID: \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" Dec 09 10:15:17 crc kubenswrapper[4824]: I1209 10:15:17.099612 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnv56\" (UniqueName: \"kubernetes.io/projected/a4e6339b-2aab-42c8-8783-5ffe08b2a026-kube-api-access-cnv56\") pod \"dnsmasq-dns-6b7bbf7cf9-lhfdn\" (UID: \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" Dec 09 10:15:17 crc kubenswrapper[4824]: I1209 10:15:17.254667 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" Dec 09 10:15:17 crc kubenswrapper[4824]: W1209 10:15:17.891072 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4e6339b_2aab_42c8_8783_5ffe08b2a026.slice/crio-f208b3a0de338f20acefd3d43724d58c6b46c52eb5854ab0a041516276a7e8c8 WatchSource:0}: Error finding container f208b3a0de338f20acefd3d43724d58c6b46c52eb5854ab0a041516276a7e8c8: Status 404 returned error can't find the container with id f208b3a0de338f20acefd3d43724d58c6b46c52eb5854ab0a041516276a7e8c8 Dec 09 10:15:17 crc kubenswrapper[4824]: I1209 10:15:17.895533 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn"] Dec 09 10:15:18 crc kubenswrapper[4824]: I1209 10:15:18.750606 4824 generic.go:334] "Generic (PLEG): container finished" podID="a4e6339b-2aab-42c8-8783-5ffe08b2a026" containerID="4cc3c3aebaf7c99d9342eb91ffcdacf5d4611de118217c6bc4ec4df04d37ce0b" exitCode=0 Dec 09 10:15:18 crc kubenswrapper[4824]: I1209 10:15:18.750670 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" event={"ID":"a4e6339b-2aab-42c8-8783-5ffe08b2a026","Type":"ContainerDied","Data":"4cc3c3aebaf7c99d9342eb91ffcdacf5d4611de118217c6bc4ec4df04d37ce0b"} Dec 09 10:15:18 crc kubenswrapper[4824]: I1209 10:15:18.751264 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" event={"ID":"a4e6339b-2aab-42c8-8783-5ffe08b2a026","Type":"ContainerStarted","Data":"f208b3a0de338f20acefd3d43724d58c6b46c52eb5854ab0a041516276a7e8c8"} Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.568563 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.655024 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/391e341b-3617-4dbf-b154-b638ba7f712a-scripts\") pod \"391e341b-3617-4dbf-b154-b638ba7f712a\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.669028 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/391e341b-3617-4dbf-b154-b638ba7f712a-scripts" (OuterVolumeSpecName: "scripts") pod "391e341b-3617-4dbf-b154-b638ba7f712a" (UID: "391e341b-3617-4dbf-b154-b638ba7f712a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.761712 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/391e341b-3617-4dbf-b154-b638ba7f712a-run-httpd\") pod \"391e341b-3617-4dbf-b154-b638ba7f712a\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.763883 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/391e341b-3617-4dbf-b154-b638ba7f712a-config-data\") pod \"391e341b-3617-4dbf-b154-b638ba7f712a\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.762076 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/391e341b-3617-4dbf-b154-b638ba7f712a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "391e341b-3617-4dbf-b154-b638ba7f712a" (UID: "391e341b-3617-4dbf-b154-b638ba7f712a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.764232 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/391e341b-3617-4dbf-b154-b638ba7f712a-combined-ca-bundle\") pod \"391e341b-3617-4dbf-b154-b638ba7f712a\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.764630 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/391e341b-3617-4dbf-b154-b638ba7f712a-log-httpd\") pod \"391e341b-3617-4dbf-b154-b638ba7f712a\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.765314 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/391e341b-3617-4dbf-b154-b638ba7f712a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "391e341b-3617-4dbf-b154-b638ba7f712a" (UID: "391e341b-3617-4dbf-b154-b638ba7f712a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.766095 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/391e341b-3617-4dbf-b154-b638ba7f712a-sg-core-conf-yaml\") pod \"391e341b-3617-4dbf-b154-b638ba7f712a\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.766253 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xtgm\" (UniqueName: \"kubernetes.io/projected/391e341b-3617-4dbf-b154-b638ba7f712a-kube-api-access-5xtgm\") pod \"391e341b-3617-4dbf-b154-b638ba7f712a\" (UID: \"391e341b-3617-4dbf-b154-b638ba7f712a\") " Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.767339 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/391e341b-3617-4dbf-b154-b638ba7f712a-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.767358 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/391e341b-3617-4dbf-b154-b638ba7f712a-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.767368 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/391e341b-3617-4dbf-b154-b638ba7f712a-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.771803 4824 generic.go:334] "Generic (PLEG): container finished" podID="391e341b-3617-4dbf-b154-b638ba7f712a" containerID="e41b24ea7f30d94712a8e4178494349c8616a0097dd1661305970dc477011c67" exitCode=0 Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.771906 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"391e341b-3617-4dbf-b154-b638ba7f712a","Type":"ContainerDied","Data":"e41b24ea7f30d94712a8e4178494349c8616a0097dd1661305970dc477011c67"} Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.771941 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"391e341b-3617-4dbf-b154-b638ba7f712a","Type":"ContainerDied","Data":"2a2ca948433846eaacf439c1f673373ed07bc59622bdfa08584a196785673e59"} Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.771962 4824 scope.go:117] "RemoveContainer" containerID="3b4fc6ee8cc316065513e16c9544a9397e005caeb346087dfa5e9025bbf2342b" Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.772563 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.773378 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/391e341b-3617-4dbf-b154-b638ba7f712a-kube-api-access-5xtgm" (OuterVolumeSpecName: "kube-api-access-5xtgm") pod "391e341b-3617-4dbf-b154-b638ba7f712a" (UID: "391e341b-3617-4dbf-b154-b638ba7f712a"). InnerVolumeSpecName "kube-api-access-5xtgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.777138 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" event={"ID":"a4e6339b-2aab-42c8-8783-5ffe08b2a026","Type":"ContainerStarted","Data":"c1ac399e73a7f3af024137814b37c0b3bf0f8b0a0fbc29b162812fc959a015ce"} Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.778397 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.804407 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" podStartSLOduration=3.8043832650000002 podStartE2EDuration="3.804383265s" podCreationTimestamp="2025-12-09 10:15:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:15:19.79884451 +0000 UTC m=+1676.133349177" watchObservedRunningTime="2025-12-09 10:15:19.804383265 +0000 UTC m=+1676.138887932" Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.809670 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/391e341b-3617-4dbf-b154-b638ba7f712a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "391e341b-3617-4dbf-b154-b638ba7f712a" (UID: "391e341b-3617-4dbf-b154-b638ba7f712a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.901515 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xtgm\" (UniqueName: \"kubernetes.io/projected/391e341b-3617-4dbf-b154-b638ba7f712a-kube-api-access-5xtgm\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.901554 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/391e341b-3617-4dbf-b154-b638ba7f712a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:19 crc kubenswrapper[4824]: I1209 10:15:19.982973 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/391e341b-3617-4dbf-b154-b638ba7f712a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "391e341b-3617-4dbf-b154-b638ba7f712a" (UID: "391e341b-3617-4dbf-b154-b638ba7f712a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.004842 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/391e341b-3617-4dbf-b154-b638ba7f712a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.007983 4824 scope.go:117] "RemoveContainer" containerID="ea40aa4f9156ace57b9937dab91e9c78a0b2d247880d0221dbb87e4fc3685676" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.025509 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/391e341b-3617-4dbf-b154-b638ba7f712a-config-data" (OuterVolumeSpecName: "config-data") pod "391e341b-3617-4dbf-b154-b638ba7f712a" (UID: "391e341b-3617-4dbf-b154-b638ba7f712a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.036205 4824 scope.go:117] "RemoveContainer" containerID="c67996f5021ab1519dc46ae411b48b0056f680f5e611031d5dc7e262e967a3da" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.072205 4824 scope.go:117] "RemoveContainer" containerID="e41b24ea7f30d94712a8e4178494349c8616a0097dd1661305970dc477011c67" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.098539 4824 scope.go:117] "RemoveContainer" containerID="3b4fc6ee8cc316065513e16c9544a9397e005caeb346087dfa5e9025bbf2342b" Dec 09 10:15:20 crc kubenswrapper[4824]: E1209 10:15:20.101064 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b4fc6ee8cc316065513e16c9544a9397e005caeb346087dfa5e9025bbf2342b\": container with ID starting with 3b4fc6ee8cc316065513e16c9544a9397e005caeb346087dfa5e9025bbf2342b not found: ID does not exist" containerID="3b4fc6ee8cc316065513e16c9544a9397e005caeb346087dfa5e9025bbf2342b" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.101112 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b4fc6ee8cc316065513e16c9544a9397e005caeb346087dfa5e9025bbf2342b"} err="failed to get container status \"3b4fc6ee8cc316065513e16c9544a9397e005caeb346087dfa5e9025bbf2342b\": rpc error: code = NotFound desc = could not find container \"3b4fc6ee8cc316065513e16c9544a9397e005caeb346087dfa5e9025bbf2342b\": container with ID starting with 3b4fc6ee8cc316065513e16c9544a9397e005caeb346087dfa5e9025bbf2342b not found: ID does not exist" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.101148 4824 scope.go:117] "RemoveContainer" containerID="ea40aa4f9156ace57b9937dab91e9c78a0b2d247880d0221dbb87e4fc3685676" Dec 09 10:15:20 crc kubenswrapper[4824]: E1209 10:15:20.101760 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea40aa4f9156ace57b9937dab91e9c78a0b2d247880d0221dbb87e4fc3685676\": container with ID starting with ea40aa4f9156ace57b9937dab91e9c78a0b2d247880d0221dbb87e4fc3685676 not found: ID does not exist" containerID="ea40aa4f9156ace57b9937dab91e9c78a0b2d247880d0221dbb87e4fc3685676" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.101805 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea40aa4f9156ace57b9937dab91e9c78a0b2d247880d0221dbb87e4fc3685676"} err="failed to get container status \"ea40aa4f9156ace57b9937dab91e9c78a0b2d247880d0221dbb87e4fc3685676\": rpc error: code = NotFound desc = could not find container \"ea40aa4f9156ace57b9937dab91e9c78a0b2d247880d0221dbb87e4fc3685676\": container with ID starting with ea40aa4f9156ace57b9937dab91e9c78a0b2d247880d0221dbb87e4fc3685676 not found: ID does not exist" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.101825 4824 scope.go:117] "RemoveContainer" containerID="c67996f5021ab1519dc46ae411b48b0056f680f5e611031d5dc7e262e967a3da" Dec 09 10:15:20 crc kubenswrapper[4824]: E1209 10:15:20.102284 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c67996f5021ab1519dc46ae411b48b0056f680f5e611031d5dc7e262e967a3da\": container with ID starting with c67996f5021ab1519dc46ae411b48b0056f680f5e611031d5dc7e262e967a3da not found: ID does not exist" containerID="c67996f5021ab1519dc46ae411b48b0056f680f5e611031d5dc7e262e967a3da" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.102361 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c67996f5021ab1519dc46ae411b48b0056f680f5e611031d5dc7e262e967a3da"} err="failed to get container status \"c67996f5021ab1519dc46ae411b48b0056f680f5e611031d5dc7e262e967a3da\": rpc error: code = NotFound desc = could not find container \"c67996f5021ab1519dc46ae411b48b0056f680f5e611031d5dc7e262e967a3da\": container with ID starting with c67996f5021ab1519dc46ae411b48b0056f680f5e611031d5dc7e262e967a3da not found: ID does not exist" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.102398 4824 scope.go:117] "RemoveContainer" containerID="e41b24ea7f30d94712a8e4178494349c8616a0097dd1661305970dc477011c67" Dec 09 10:15:20 crc kubenswrapper[4824]: E1209 10:15:20.102802 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e41b24ea7f30d94712a8e4178494349c8616a0097dd1661305970dc477011c67\": container with ID starting with e41b24ea7f30d94712a8e4178494349c8616a0097dd1661305970dc477011c67 not found: ID does not exist" containerID="e41b24ea7f30d94712a8e4178494349c8616a0097dd1661305970dc477011c67" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.102853 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e41b24ea7f30d94712a8e4178494349c8616a0097dd1661305970dc477011c67"} err="failed to get container status \"e41b24ea7f30d94712a8e4178494349c8616a0097dd1661305970dc477011c67\": rpc error: code = NotFound desc = could not find container \"e41b24ea7f30d94712a8e4178494349c8616a0097dd1661305970dc477011c67\": container with ID starting with e41b24ea7f30d94712a8e4178494349c8616a0097dd1661305970dc477011c67 not found: ID does not exist" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.110755 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/391e341b-3617-4dbf-b154-b638ba7f712a-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.127258 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.141730 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.173518 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:15:20 crc kubenswrapper[4824]: E1209 10:15:20.174464 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="391e341b-3617-4dbf-b154-b638ba7f712a" containerName="ceilometer-notification-agent" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.174599 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="391e341b-3617-4dbf-b154-b638ba7f712a" containerName="ceilometer-notification-agent" Dec 09 10:15:20 crc kubenswrapper[4824]: E1209 10:15:20.174690 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="391e341b-3617-4dbf-b154-b638ba7f712a" containerName="sg-core" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.174774 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="391e341b-3617-4dbf-b154-b638ba7f712a" containerName="sg-core" Dec 09 10:15:20 crc kubenswrapper[4824]: E1209 10:15:20.174904 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="391e341b-3617-4dbf-b154-b638ba7f712a" containerName="proxy-httpd" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.174978 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="391e341b-3617-4dbf-b154-b638ba7f712a" containerName="proxy-httpd" Dec 09 10:15:20 crc kubenswrapper[4824]: E1209 10:15:20.175085 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="391e341b-3617-4dbf-b154-b638ba7f712a" containerName="ceilometer-central-agent" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.175159 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="391e341b-3617-4dbf-b154-b638ba7f712a" containerName="ceilometer-central-agent" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.175522 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="391e341b-3617-4dbf-b154-b638ba7f712a" containerName="sg-core" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.175600 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="391e341b-3617-4dbf-b154-b638ba7f712a" containerName="ceilometer-central-agent" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.175702 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="391e341b-3617-4dbf-b154-b638ba7f712a" containerName="ceilometer-notification-agent" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.175766 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="391e341b-3617-4dbf-b154-b638ba7f712a" containerName="proxy-httpd" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.197257 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.197647 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.203272 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.203528 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.204060 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.226461 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.226608 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-scripts\") pod \"ceilometer-0\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.226628 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d254951a-1dd1-4b09-8cfa-e765853cb5e9-log-httpd\") pod \"ceilometer-0\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.226707 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d254951a-1dd1-4b09-8cfa-e765853cb5e9-run-httpd\") pod \"ceilometer-0\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.226759 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dlhv\" (UniqueName: \"kubernetes.io/projected/d254951a-1dd1-4b09-8cfa-e765853cb5e9-kube-api-access-9dlhv\") pod \"ceilometer-0\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.226798 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-config-data\") pod \"ceilometer-0\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.226939 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.227023 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.355879 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.356297 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.356687 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.356975 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-scripts\") pod \"ceilometer-0\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.357086 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d254951a-1dd1-4b09-8cfa-e765853cb5e9-log-httpd\") pod \"ceilometer-0\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.357310 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d254951a-1dd1-4b09-8cfa-e765853cb5e9-run-httpd\") pod \"ceilometer-0\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.357499 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dlhv\" (UniqueName: \"kubernetes.io/projected/d254951a-1dd1-4b09-8cfa-e765853cb5e9-kube-api-access-9dlhv\") pod \"ceilometer-0\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.357611 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-config-data\") pod \"ceilometer-0\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.366843 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d254951a-1dd1-4b09-8cfa-e765853cb5e9-run-httpd\") pod \"ceilometer-0\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.368120 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d254951a-1dd1-4b09-8cfa-e765853cb5e9-log-httpd\") pod \"ceilometer-0\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.368261 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.368316 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.368683 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-config-data\") pod \"ceilometer-0\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.368710 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.369261 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-scripts\") pod \"ceilometer-0\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.389695 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dlhv\" (UniqueName: \"kubernetes.io/projected/d254951a-1dd1-4b09-8cfa-e765853cb5e9-kube-api-access-9dlhv\") pod \"ceilometer-0\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.451191 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.451625 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="74053563-b115-4fb2-9433-16e81288df8d" containerName="nova-api-log" containerID="cri-o://1dd6986809d86b0375eba3e70ea42bde9ef897c1c0331e219811b0dcc6c09a73" gracePeriod=30 Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.451908 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="74053563-b115-4fb2-9433-16e81288df8d" containerName="nova-api-api" containerID="cri-o://4b6d495e7633ad7352a5222410c3cfc1f99d6b25ec2d79c143b79fe408a02c74" gracePeriod=30 Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.546667 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.746218 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.781329 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.850918 4824 generic.go:334] "Generic (PLEG): container finished" podID="74053563-b115-4fb2-9433-16e81288df8d" containerID="1dd6986809d86b0375eba3e70ea42bde9ef897c1c0331e219811b0dcc6c09a73" exitCode=143 Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.851000 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"74053563-b115-4fb2-9433-16e81288df8d","Type":"ContainerDied","Data":"1dd6986809d86b0375eba3e70ea42bde9ef897c1c0331e219811b0dcc6c09a73"} Dec 09 10:15:20 crc kubenswrapper[4824]: I1209 10:15:20.973224 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 09 10:15:21 crc kubenswrapper[4824]: I1209 10:15:21.280157 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:15:21 crc kubenswrapper[4824]: I1209 10:15:21.536710 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-282jv"] Dec 09 10:15:21 crc kubenswrapper[4824]: I1209 10:15:21.539254 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-282jv" Dec 09 10:15:21 crc kubenswrapper[4824]: I1209 10:15:21.543755 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 09 10:15:21 crc kubenswrapper[4824]: I1209 10:15:21.558652 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 09 10:15:21 crc kubenswrapper[4824]: I1209 10:15:21.592854 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-282jv"] Dec 09 10:15:21 crc kubenswrapper[4824]: I1209 10:15:21.668595 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a34cbc14-a8b4-4993-89ba-b36f8a9a4065-scripts\") pod \"nova-cell1-cell-mapping-282jv\" (UID: \"a34cbc14-a8b4-4993-89ba-b36f8a9a4065\") " pod="openstack/nova-cell1-cell-mapping-282jv" Dec 09 10:15:21 crc kubenswrapper[4824]: I1209 10:15:21.668832 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfk6s\" (UniqueName: \"kubernetes.io/projected/a34cbc14-a8b4-4993-89ba-b36f8a9a4065-kube-api-access-zfk6s\") pod \"nova-cell1-cell-mapping-282jv\" (UID: \"a34cbc14-a8b4-4993-89ba-b36f8a9a4065\") " pod="openstack/nova-cell1-cell-mapping-282jv" Dec 09 10:15:21 crc kubenswrapper[4824]: I1209 10:15:21.669751 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a34cbc14-a8b4-4993-89ba-b36f8a9a4065-config-data\") pod \"nova-cell1-cell-mapping-282jv\" (UID: \"a34cbc14-a8b4-4993-89ba-b36f8a9a4065\") " pod="openstack/nova-cell1-cell-mapping-282jv" Dec 09 10:15:21 crc kubenswrapper[4824]: I1209 10:15:21.669894 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a34cbc14-a8b4-4993-89ba-b36f8a9a4065-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-282jv\" (UID: \"a34cbc14-a8b4-4993-89ba-b36f8a9a4065\") " pod="openstack/nova-cell1-cell-mapping-282jv" Dec 09 10:15:21 crc kubenswrapper[4824]: I1209 10:15:21.772591 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a34cbc14-a8b4-4993-89ba-b36f8a9a4065-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-282jv\" (UID: \"a34cbc14-a8b4-4993-89ba-b36f8a9a4065\") " pod="openstack/nova-cell1-cell-mapping-282jv" Dec 09 10:15:21 crc kubenswrapper[4824]: I1209 10:15:21.772797 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a34cbc14-a8b4-4993-89ba-b36f8a9a4065-scripts\") pod \"nova-cell1-cell-mapping-282jv\" (UID: \"a34cbc14-a8b4-4993-89ba-b36f8a9a4065\") " pod="openstack/nova-cell1-cell-mapping-282jv" Dec 09 10:15:21 crc kubenswrapper[4824]: I1209 10:15:21.772875 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfk6s\" (UniqueName: \"kubernetes.io/projected/a34cbc14-a8b4-4993-89ba-b36f8a9a4065-kube-api-access-zfk6s\") pod \"nova-cell1-cell-mapping-282jv\" (UID: \"a34cbc14-a8b4-4993-89ba-b36f8a9a4065\") " pod="openstack/nova-cell1-cell-mapping-282jv" Dec 09 10:15:21 crc kubenswrapper[4824]: I1209 10:15:21.773138 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a34cbc14-a8b4-4993-89ba-b36f8a9a4065-config-data\") pod \"nova-cell1-cell-mapping-282jv\" (UID: \"a34cbc14-a8b4-4993-89ba-b36f8a9a4065\") " pod="openstack/nova-cell1-cell-mapping-282jv" Dec 09 10:15:21 crc kubenswrapper[4824]: I1209 10:15:21.780670 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a34cbc14-a8b4-4993-89ba-b36f8a9a4065-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-282jv\" (UID: \"a34cbc14-a8b4-4993-89ba-b36f8a9a4065\") " pod="openstack/nova-cell1-cell-mapping-282jv" Dec 09 10:15:21 crc kubenswrapper[4824]: I1209 10:15:21.792081 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a34cbc14-a8b4-4993-89ba-b36f8a9a4065-config-data\") pod \"nova-cell1-cell-mapping-282jv\" (UID: \"a34cbc14-a8b4-4993-89ba-b36f8a9a4065\") " pod="openstack/nova-cell1-cell-mapping-282jv" Dec 09 10:15:21 crc kubenswrapper[4824]: I1209 10:15:21.794629 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a34cbc14-a8b4-4993-89ba-b36f8a9a4065-scripts\") pod \"nova-cell1-cell-mapping-282jv\" (UID: \"a34cbc14-a8b4-4993-89ba-b36f8a9a4065\") " pod="openstack/nova-cell1-cell-mapping-282jv" Dec 09 10:15:21 crc kubenswrapper[4824]: I1209 10:15:21.794853 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfk6s\" (UniqueName: \"kubernetes.io/projected/a34cbc14-a8b4-4993-89ba-b36f8a9a4065-kube-api-access-zfk6s\") pod \"nova-cell1-cell-mapping-282jv\" (UID: \"a34cbc14-a8b4-4993-89ba-b36f8a9a4065\") " pod="openstack/nova-cell1-cell-mapping-282jv" Dec 09 10:15:21 crc kubenswrapper[4824]: I1209 10:15:21.884643 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-282jv" Dec 09 10:15:22 crc kubenswrapper[4824]: I1209 10:15:22.043971 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="391e341b-3617-4dbf-b154-b638ba7f712a" path="/var/lib/kubelet/pods/391e341b-3617-4dbf-b154-b638ba7f712a/volumes" Dec 09 10:15:22 crc kubenswrapper[4824]: I1209 10:15:22.045525 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d254951a-1dd1-4b09-8cfa-e765853cb5e9","Type":"ContainerStarted","Data":"4f8718a387b02f1a62c3272b0b6fc52b6dd12ac62f11c441842290a0d4102a51"} Dec 09 10:15:22 crc kubenswrapper[4824]: I1209 10:15:22.078448 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 09 10:15:22 crc kubenswrapper[4824]: I1209 10:15:22.246031 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:15:22 crc kubenswrapper[4824]: I1209 10:15:22.723562 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-282jv"] Dec 09 10:15:23 crc kubenswrapper[4824]: I1209 10:15:23.099182 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d254951a-1dd1-4b09-8cfa-e765853cb5e9","Type":"ContainerStarted","Data":"1cd4211ca88d7f9721e9b7c47366ab131e71cdd2d1bc02648cdec6c0c0719677"} Dec 09 10:15:23 crc kubenswrapper[4824]: I1209 10:15:23.107727 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-282jv" event={"ID":"a34cbc14-a8b4-4993-89ba-b36f8a9a4065","Type":"ContainerStarted","Data":"f7e64ca96567bb5d056db62dd887e885a3cb90a04ca0609a4db893c500c19962"} Dec 09 10:15:24 crc kubenswrapper[4824]: I1209 10:15:24.191113 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d254951a-1dd1-4b09-8cfa-e765853cb5e9","Type":"ContainerStarted","Data":"aa021504dd953ba53062645025462a4f01abb00b32816d01f0196aaa2ade5b7d"} Dec 09 10:15:24 crc kubenswrapper[4824]: I1209 10:15:24.200940 4824 generic.go:334] "Generic (PLEG): container finished" podID="74053563-b115-4fb2-9433-16e81288df8d" containerID="4b6d495e7633ad7352a5222410c3cfc1f99d6b25ec2d79c143b79fe408a02c74" exitCode=0 Dec 09 10:15:24 crc kubenswrapper[4824]: I1209 10:15:24.201002 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"74053563-b115-4fb2-9433-16e81288df8d","Type":"ContainerDied","Data":"4b6d495e7633ad7352a5222410c3cfc1f99d6b25ec2d79c143b79fe408a02c74"} Dec 09 10:15:24 crc kubenswrapper[4824]: I1209 10:15:24.212033 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-282jv" event={"ID":"a34cbc14-a8b4-4993-89ba-b36f8a9a4065","Type":"ContainerStarted","Data":"f5d20ae0706a64448caf4d1146e14931c247faafec1c1471ff9e5c97c20a91a4"} Dec 09 10:15:24 crc kubenswrapper[4824]: I1209 10:15:24.241470 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-282jv" podStartSLOduration=3.241450182 podStartE2EDuration="3.241450182s" podCreationTimestamp="2025-12-09 10:15:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:15:24.239095298 +0000 UTC m=+1680.573599965" watchObservedRunningTime="2025-12-09 10:15:24.241450182 +0000 UTC m=+1680.575954849" Dec 09 10:15:24 crc kubenswrapper[4824]: I1209 10:15:24.500316 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 10:15:24 crc kubenswrapper[4824]: I1209 10:15:24.637842 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74053563-b115-4fb2-9433-16e81288df8d-combined-ca-bundle\") pod \"74053563-b115-4fb2-9433-16e81288df8d\" (UID: \"74053563-b115-4fb2-9433-16e81288df8d\") " Dec 09 10:15:24 crc kubenswrapper[4824]: I1209 10:15:24.638082 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lgbh\" (UniqueName: \"kubernetes.io/projected/74053563-b115-4fb2-9433-16e81288df8d-kube-api-access-4lgbh\") pod \"74053563-b115-4fb2-9433-16e81288df8d\" (UID: \"74053563-b115-4fb2-9433-16e81288df8d\") " Dec 09 10:15:24 crc kubenswrapper[4824]: I1209 10:15:24.638300 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74053563-b115-4fb2-9433-16e81288df8d-logs\") pod \"74053563-b115-4fb2-9433-16e81288df8d\" (UID: \"74053563-b115-4fb2-9433-16e81288df8d\") " Dec 09 10:15:24 crc kubenswrapper[4824]: I1209 10:15:24.638401 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74053563-b115-4fb2-9433-16e81288df8d-config-data\") pod \"74053563-b115-4fb2-9433-16e81288df8d\" (UID: \"74053563-b115-4fb2-9433-16e81288df8d\") " Dec 09 10:15:24 crc kubenswrapper[4824]: I1209 10:15:24.638862 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74053563-b115-4fb2-9433-16e81288df8d-logs" (OuterVolumeSpecName: "logs") pod "74053563-b115-4fb2-9433-16e81288df8d" (UID: "74053563-b115-4fb2-9433-16e81288df8d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:15:24 crc kubenswrapper[4824]: I1209 10:15:24.639657 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74053563-b115-4fb2-9433-16e81288df8d-logs\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:24 crc kubenswrapper[4824]: I1209 10:15:24.644146 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74053563-b115-4fb2-9433-16e81288df8d-kube-api-access-4lgbh" (OuterVolumeSpecName: "kube-api-access-4lgbh") pod "74053563-b115-4fb2-9433-16e81288df8d" (UID: "74053563-b115-4fb2-9433-16e81288df8d"). InnerVolumeSpecName "kube-api-access-4lgbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:15:24 crc kubenswrapper[4824]: I1209 10:15:24.715965 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74053563-b115-4fb2-9433-16e81288df8d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74053563-b115-4fb2-9433-16e81288df8d" (UID: "74053563-b115-4fb2-9433-16e81288df8d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:24 crc kubenswrapper[4824]: I1209 10:15:24.719191 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74053563-b115-4fb2-9433-16e81288df8d-config-data" (OuterVolumeSpecName: "config-data") pod "74053563-b115-4fb2-9433-16e81288df8d" (UID: "74053563-b115-4fb2-9433-16e81288df8d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:24 crc kubenswrapper[4824]: I1209 10:15:24.741615 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lgbh\" (UniqueName: \"kubernetes.io/projected/74053563-b115-4fb2-9433-16e81288df8d-kube-api-access-4lgbh\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:24 crc kubenswrapper[4824]: I1209 10:15:24.741677 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74053563-b115-4fb2-9433-16e81288df8d-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:24 crc kubenswrapper[4824]: I1209 10:15:24.741688 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74053563-b115-4fb2-9433-16e81288df8d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.232162 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d254951a-1dd1-4b09-8cfa-e765853cb5e9","Type":"ContainerStarted","Data":"6ecc53fd2da09ec596a0864786ec17e2f907598c43db35eabbe9f964824bb708"} Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.234892 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"74053563-b115-4fb2-9433-16e81288df8d","Type":"ContainerDied","Data":"26dfbca4840c6bb147000031c2cd4d7b5551267e552f941277a67648603422d0"} Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.234957 4824 scope.go:117] "RemoveContainer" containerID="4b6d495e7633ad7352a5222410c3cfc1f99d6b25ec2d79c143b79fe408a02c74" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.235926 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.296632 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.299645 4824 scope.go:117] "RemoveContainer" containerID="1dd6986809d86b0375eba3e70ea42bde9ef897c1c0331e219811b0dcc6c09a73" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.321719 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.349027 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 09 10:15:25 crc kubenswrapper[4824]: E1209 10:15:25.349863 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74053563-b115-4fb2-9433-16e81288df8d" containerName="nova-api-api" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.349885 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="74053563-b115-4fb2-9433-16e81288df8d" containerName="nova-api-api" Dec 09 10:15:25 crc kubenswrapper[4824]: E1209 10:15:25.349906 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74053563-b115-4fb2-9433-16e81288df8d" containerName="nova-api-log" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.349913 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="74053563-b115-4fb2-9433-16e81288df8d" containerName="nova-api-log" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.350265 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="74053563-b115-4fb2-9433-16e81288df8d" containerName="nova-api-log" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.350295 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="74053563-b115-4fb2-9433-16e81288df8d" containerName="nova-api-api" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.352218 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.359119 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.359361 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.359488 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.361812 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.471314 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhg46\" (UniqueName: \"kubernetes.io/projected/21c56d0b-49a8-4fdd-964f-93a14d1dd003-kube-api-access-rhg46\") pod \"nova-api-0\" (UID: \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\") " pod="openstack/nova-api-0" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.471393 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21c56d0b-49a8-4fdd-964f-93a14d1dd003-config-data\") pod \"nova-api-0\" (UID: \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\") " pod="openstack/nova-api-0" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.471416 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21c56d0b-49a8-4fdd-964f-93a14d1dd003-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\") " pod="openstack/nova-api-0" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.471543 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21c56d0b-49a8-4fdd-964f-93a14d1dd003-internal-tls-certs\") pod \"nova-api-0\" (UID: \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\") " pod="openstack/nova-api-0" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.471623 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21c56d0b-49a8-4fdd-964f-93a14d1dd003-public-tls-certs\") pod \"nova-api-0\" (UID: \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\") " pod="openstack/nova-api-0" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.471846 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21c56d0b-49a8-4fdd-964f-93a14d1dd003-logs\") pod \"nova-api-0\" (UID: \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\") " pod="openstack/nova-api-0" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.613606 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21c56d0b-49a8-4fdd-964f-93a14d1dd003-logs\") pod \"nova-api-0\" (UID: \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\") " pod="openstack/nova-api-0" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.614153 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21c56d0b-49a8-4fdd-964f-93a14d1dd003-logs\") pod \"nova-api-0\" (UID: \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\") " pod="openstack/nova-api-0" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.614228 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhg46\" (UniqueName: \"kubernetes.io/projected/21c56d0b-49a8-4fdd-964f-93a14d1dd003-kube-api-access-rhg46\") pod \"nova-api-0\" (UID: \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\") " pod="openstack/nova-api-0" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.614370 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21c56d0b-49a8-4fdd-964f-93a14d1dd003-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\") " pod="openstack/nova-api-0" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.614409 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21c56d0b-49a8-4fdd-964f-93a14d1dd003-config-data\") pod \"nova-api-0\" (UID: \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\") " pod="openstack/nova-api-0" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.616611 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21c56d0b-49a8-4fdd-964f-93a14d1dd003-internal-tls-certs\") pod \"nova-api-0\" (UID: \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\") " pod="openstack/nova-api-0" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.616712 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21c56d0b-49a8-4fdd-964f-93a14d1dd003-public-tls-certs\") pod \"nova-api-0\" (UID: \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\") " pod="openstack/nova-api-0" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.621875 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21c56d0b-49a8-4fdd-964f-93a14d1dd003-public-tls-certs\") pod \"nova-api-0\" (UID: \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\") " pod="openstack/nova-api-0" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.622273 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21c56d0b-49a8-4fdd-964f-93a14d1dd003-config-data\") pod \"nova-api-0\" (UID: \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\") " pod="openstack/nova-api-0" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.622389 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21c56d0b-49a8-4fdd-964f-93a14d1dd003-internal-tls-certs\") pod \"nova-api-0\" (UID: \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\") " pod="openstack/nova-api-0" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.624364 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21c56d0b-49a8-4fdd-964f-93a14d1dd003-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\") " pod="openstack/nova-api-0" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.654251 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhg46\" (UniqueName: \"kubernetes.io/projected/21c56d0b-49a8-4fdd-964f-93a14d1dd003-kube-api-access-rhg46\") pod \"nova-api-0\" (UID: \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\") " pod="openstack/nova-api-0" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.684520 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 10:15:25 crc kubenswrapper[4824]: I1209 10:15:25.970397 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74053563-b115-4fb2-9433-16e81288df8d" path="/var/lib/kubelet/pods/74053563-b115-4fb2-9433-16e81288df8d/volumes" Dec 09 10:15:26 crc kubenswrapper[4824]: I1209 10:15:26.295485 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d254951a-1dd1-4b09-8cfa-e765853cb5e9","Type":"ContainerStarted","Data":"a2f5f62a717870a67acb6dd5fc736f75d1a2b516521177079b70d23cdbe5ce0f"} Dec 09 10:15:26 crc kubenswrapper[4824]: I1209 10:15:26.296110 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d254951a-1dd1-4b09-8cfa-e765853cb5e9" containerName="ceilometer-central-agent" containerID="cri-o://1cd4211ca88d7f9721e9b7c47366ab131e71cdd2d1bc02648cdec6c0c0719677" gracePeriod=30 Dec 09 10:15:26 crc kubenswrapper[4824]: I1209 10:15:26.296421 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d254951a-1dd1-4b09-8cfa-e765853cb5e9" containerName="sg-core" containerID="cri-o://6ecc53fd2da09ec596a0864786ec17e2f907598c43db35eabbe9f964824bb708" gracePeriod=30 Dec 09 10:15:26 crc kubenswrapper[4824]: I1209 10:15:26.296446 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d254951a-1dd1-4b09-8cfa-e765853cb5e9" containerName="ceilometer-notification-agent" containerID="cri-o://aa021504dd953ba53062645025462a4f01abb00b32816d01f0196aaa2ade5b7d" gracePeriod=30 Dec 09 10:15:26 crc kubenswrapper[4824]: I1209 10:15:26.296381 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d254951a-1dd1-4b09-8cfa-e765853cb5e9" containerName="proxy-httpd" containerID="cri-o://a2f5f62a717870a67acb6dd5fc736f75d1a2b516521177079b70d23cdbe5ce0f" gracePeriod=30 Dec 09 10:15:26 crc kubenswrapper[4824]: I1209 10:15:26.344265 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 10:15:26 crc kubenswrapper[4824]: I1209 10:15:26.356913 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.16648053 podStartE2EDuration="6.356868333s" podCreationTimestamp="2025-12-09 10:15:20 +0000 UTC" firstStartedPulling="2025-12-09 10:15:21.303998788 +0000 UTC m=+1677.638503455" lastFinishedPulling="2025-12-09 10:15:25.494386591 +0000 UTC m=+1681.828891258" observedRunningTime="2025-12-09 10:15:26.341014892 +0000 UTC m=+1682.675519559" watchObservedRunningTime="2025-12-09 10:15:26.356868333 +0000 UTC m=+1682.691373000" Dec 09 10:15:27 crc kubenswrapper[4824]: I1209 10:15:27.256753 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" Dec 09 10:15:27 crc kubenswrapper[4824]: I1209 10:15:27.344122 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"21c56d0b-49a8-4fdd-964f-93a14d1dd003","Type":"ContainerStarted","Data":"32dfc962e7123d7f0622108b65ca08b023a7eafbc684703a1b52ed512f92f127"} Dec 09 10:15:27 crc kubenswrapper[4824]: I1209 10:15:27.344174 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"21c56d0b-49a8-4fdd-964f-93a14d1dd003","Type":"ContainerStarted","Data":"1b314689dbbbace6f1ed19df9e216a45cfac2b1150ebe53f83fc407cff1e7f96"} Dec 09 10:15:27 crc kubenswrapper[4824]: I1209 10:15:27.344190 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"21c56d0b-49a8-4fdd-964f-93a14d1dd003","Type":"ContainerStarted","Data":"c79dcecaa53a8c479cf25b09dab2cbc809af4579142844d46d823400baa3763b"} Dec 09 10:15:27 crc kubenswrapper[4824]: I1209 10:15:27.372973 4824 generic.go:334] "Generic (PLEG): container finished" podID="d254951a-1dd1-4b09-8cfa-e765853cb5e9" containerID="6ecc53fd2da09ec596a0864786ec17e2f907598c43db35eabbe9f964824bb708" exitCode=2 Dec 09 10:15:27 crc kubenswrapper[4824]: I1209 10:15:27.373012 4824 generic.go:334] "Generic (PLEG): container finished" podID="d254951a-1dd1-4b09-8cfa-e765853cb5e9" containerID="aa021504dd953ba53062645025462a4f01abb00b32816d01f0196aaa2ade5b7d" exitCode=0 Dec 09 10:15:27 crc kubenswrapper[4824]: I1209 10:15:27.373038 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d254951a-1dd1-4b09-8cfa-e765853cb5e9","Type":"ContainerDied","Data":"6ecc53fd2da09ec596a0864786ec17e2f907598c43db35eabbe9f964824bb708"} Dec 09 10:15:27 crc kubenswrapper[4824]: I1209 10:15:27.373069 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d254951a-1dd1-4b09-8cfa-e765853cb5e9","Type":"ContainerDied","Data":"aa021504dd953ba53062645025462a4f01abb00b32816d01f0196aaa2ade5b7d"} Dec 09 10:15:27 crc kubenswrapper[4824]: I1209 10:15:27.374301 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-qcgm4"] Dec 09 10:15:27 crc kubenswrapper[4824]: I1209 10:15:27.377135 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" podUID="b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6" containerName="dnsmasq-dns" containerID="cri-o://b699134baa3469d213951ac530d09148ea90fdf79e2fe5574dc7eb9184ca9b37" gracePeriod=10 Dec 09 10:15:27 crc kubenswrapper[4824]: I1209 10:15:27.407514 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.407485049 podStartE2EDuration="2.407485049s" podCreationTimestamp="2025-12-09 10:15:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:15:27.403057719 +0000 UTC m=+1683.737562396" watchObservedRunningTime="2025-12-09 10:15:27.407485049 +0000 UTC m=+1683.741989726" Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.140096 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.349299 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-config\") pod \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\" (UID: \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\") " Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.349513 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9nj5\" (UniqueName: \"kubernetes.io/projected/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-kube-api-access-x9nj5\") pod \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\" (UID: \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\") " Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.349709 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-ovsdbserver-sb\") pod \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\" (UID: \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\") " Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.349742 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-dns-swift-storage-0\") pod \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\" (UID: \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\") " Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.349803 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-ovsdbserver-nb\") pod \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\" (UID: \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\") " Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.349877 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-dns-svc\") pod \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\" (UID: \"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6\") " Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.392767 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-kube-api-access-x9nj5" (OuterVolumeSpecName: "kube-api-access-x9nj5") pod "b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6" (UID: "b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6"). InnerVolumeSpecName "kube-api-access-x9nj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.453596 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9nj5\" (UniqueName: \"kubernetes.io/projected/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-kube-api-access-x9nj5\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.502238 4824 generic.go:334] "Generic (PLEG): container finished" podID="b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6" containerID="b699134baa3469d213951ac530d09148ea90fdf79e2fe5574dc7eb9184ca9b37" exitCode=0 Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.503695 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6" (UID: "b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.503807 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.504321 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" event={"ID":"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6","Type":"ContainerDied","Data":"b699134baa3469d213951ac530d09148ea90fdf79e2fe5574dc7eb9184ca9b37"} Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.504370 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-qcgm4" event={"ID":"b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6","Type":"ContainerDied","Data":"cdec60c4257e65e9647e5213309d6ec9812bacba891b84c2c517eff535223385"} Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.504393 4824 scope.go:117] "RemoveContainer" containerID="b699134baa3469d213951ac530d09148ea90fdf79e2fe5574dc7eb9184ca9b37" Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.525390 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6" (UID: "b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.538290 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6" (UID: "b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.556859 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.556897 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.556906 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.637412 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-config" (OuterVolumeSpecName: "config") pod "b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6" (UID: "b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.659898 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.695618 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6" (UID: "b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.837834 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:28 crc kubenswrapper[4824]: I1209 10:15:28.981977 4824 scope.go:117] "RemoveContainer" containerID="1206eefde90c3313178d69b7b139b9d2fae3549a611ba3ca14e1378fa5e37c78" Dec 09 10:15:29 crc kubenswrapper[4824]: I1209 10:15:29.017888 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-qcgm4"] Dec 09 10:15:29 crc kubenswrapper[4824]: I1209 10:15:29.042774 4824 scope.go:117] "RemoveContainer" containerID="b699134baa3469d213951ac530d09148ea90fdf79e2fe5574dc7eb9184ca9b37" Dec 09 10:15:29 crc kubenswrapper[4824]: I1209 10:15:29.043176 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-qcgm4"] Dec 09 10:15:29 crc kubenswrapper[4824]: E1209 10:15:29.054678 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b699134baa3469d213951ac530d09148ea90fdf79e2fe5574dc7eb9184ca9b37\": container with ID starting with b699134baa3469d213951ac530d09148ea90fdf79e2fe5574dc7eb9184ca9b37 not found: ID does not exist" containerID="b699134baa3469d213951ac530d09148ea90fdf79e2fe5574dc7eb9184ca9b37" Dec 09 10:15:29 crc kubenswrapper[4824]: I1209 10:15:29.054990 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b699134baa3469d213951ac530d09148ea90fdf79e2fe5574dc7eb9184ca9b37"} err="failed to get container status \"b699134baa3469d213951ac530d09148ea90fdf79e2fe5574dc7eb9184ca9b37\": rpc error: code = NotFound desc = could not find container \"b699134baa3469d213951ac530d09148ea90fdf79e2fe5574dc7eb9184ca9b37\": container with ID starting with b699134baa3469d213951ac530d09148ea90fdf79e2fe5574dc7eb9184ca9b37 not found: ID does not exist" Dec 09 10:15:29 crc kubenswrapper[4824]: I1209 10:15:29.055025 4824 scope.go:117] "RemoveContainer" containerID="1206eefde90c3313178d69b7b139b9d2fae3549a611ba3ca14e1378fa5e37c78" Dec 09 10:15:29 crc kubenswrapper[4824]: E1209 10:15:29.055501 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1206eefde90c3313178d69b7b139b9d2fae3549a611ba3ca14e1378fa5e37c78\": container with ID starting with 1206eefde90c3313178d69b7b139b9d2fae3549a611ba3ca14e1378fa5e37c78 not found: ID does not exist" containerID="1206eefde90c3313178d69b7b139b9d2fae3549a611ba3ca14e1378fa5e37c78" Dec 09 10:15:29 crc kubenswrapper[4824]: I1209 10:15:29.055540 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1206eefde90c3313178d69b7b139b9d2fae3549a611ba3ca14e1378fa5e37c78"} err="failed to get container status \"1206eefde90c3313178d69b7b139b9d2fae3549a611ba3ca14e1378fa5e37c78\": rpc error: code = NotFound desc = could not find container \"1206eefde90c3313178d69b7b139b9d2fae3549a611ba3ca14e1378fa5e37c78\": container with ID starting with 1206eefde90c3313178d69b7b139b9d2fae3549a611ba3ca14e1378fa5e37c78 not found: ID does not exist" Dec 09 10:15:29 crc kubenswrapper[4824]: E1209 10:15:29.696840 4824 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/55fbb64abefb902ed865c19450931e3838777b9f9a8ebf97ebece28f46c92de9/diff" to get inode usage: stat /var/lib/containers/storage/overlay/55fbb64abefb902ed865c19450931e3838777b9f9a8ebf97ebece28f46c92de9/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_dnsmasq-dns-9b86998b5-qcgm4_b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6/dnsmasq-dns/0.log" to get inode usage: stat /var/log/pods/openstack_dnsmasq-dns-9b86998b5-qcgm4_b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6/dnsmasq-dns/0.log: no such file or directory Dec 09 10:15:29 crc kubenswrapper[4824]: I1209 10:15:29.925151 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6" path="/var/lib/kubelet/pods/b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6/volumes" Dec 09 10:15:31 crc kubenswrapper[4824]: I1209 10:15:31.549190 4824 generic.go:334] "Generic (PLEG): container finished" podID="a34cbc14-a8b4-4993-89ba-b36f8a9a4065" containerID="f5d20ae0706a64448caf4d1146e14931c247faafec1c1471ff9e5c97c20a91a4" exitCode=0 Dec 09 10:15:31 crc kubenswrapper[4824]: I1209 10:15:31.549270 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-282jv" event={"ID":"a34cbc14-a8b4-4993-89ba-b36f8a9a4065","Type":"ContainerDied","Data":"f5d20ae0706a64448caf4d1146e14931c247faafec1c1471ff9e5c97c20a91a4"} Dec 09 10:15:32 crc kubenswrapper[4824]: I1209 10:15:32.564458 4824 generic.go:334] "Generic (PLEG): container finished" podID="d254951a-1dd1-4b09-8cfa-e765853cb5e9" containerID="1cd4211ca88d7f9721e9b7c47366ab131e71cdd2d1bc02648cdec6c0c0719677" exitCode=0 Dec 09 10:15:32 crc kubenswrapper[4824]: I1209 10:15:32.565011 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d254951a-1dd1-4b09-8cfa-e765853cb5e9","Type":"ContainerDied","Data":"1cd4211ca88d7f9721e9b7c47366ab131e71cdd2d1bc02648cdec6c0c0719677"} Dec 09 10:15:32 crc kubenswrapper[4824]: I1209 10:15:32.860754 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:15:32 crc kubenswrapper[4824]: I1209 10:15:32.861141 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:15:32 crc kubenswrapper[4824]: I1209 10:15:32.861192 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 10:15:32 crc kubenswrapper[4824]: I1209 10:15:32.862166 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1"} pod="openshift-machine-config-operator/machine-config-daemon-dth8x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 10:15:32 crc kubenswrapper[4824]: I1209 10:15:32.862235 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" containerID="cri-o://7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" gracePeriod=600 Dec 09 10:15:32 crc kubenswrapper[4824]: E1209 10:15:32.989295 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:15:33 crc kubenswrapper[4824]: I1209 10:15:33.070371 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-282jv" Dec 09 10:15:33 crc kubenswrapper[4824]: I1209 10:15:33.181264 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfk6s\" (UniqueName: \"kubernetes.io/projected/a34cbc14-a8b4-4993-89ba-b36f8a9a4065-kube-api-access-zfk6s\") pod \"a34cbc14-a8b4-4993-89ba-b36f8a9a4065\" (UID: \"a34cbc14-a8b4-4993-89ba-b36f8a9a4065\") " Dec 09 10:15:33 crc kubenswrapper[4824]: I1209 10:15:33.181318 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a34cbc14-a8b4-4993-89ba-b36f8a9a4065-combined-ca-bundle\") pod \"a34cbc14-a8b4-4993-89ba-b36f8a9a4065\" (UID: \"a34cbc14-a8b4-4993-89ba-b36f8a9a4065\") " Dec 09 10:15:33 crc kubenswrapper[4824]: I1209 10:15:33.181597 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a34cbc14-a8b4-4993-89ba-b36f8a9a4065-config-data\") pod \"a34cbc14-a8b4-4993-89ba-b36f8a9a4065\" (UID: \"a34cbc14-a8b4-4993-89ba-b36f8a9a4065\") " Dec 09 10:15:33 crc kubenswrapper[4824]: I1209 10:15:33.181641 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a34cbc14-a8b4-4993-89ba-b36f8a9a4065-scripts\") pod \"a34cbc14-a8b4-4993-89ba-b36f8a9a4065\" (UID: \"a34cbc14-a8b4-4993-89ba-b36f8a9a4065\") " Dec 09 10:15:33 crc kubenswrapper[4824]: I1209 10:15:33.198120 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a34cbc14-a8b4-4993-89ba-b36f8a9a4065-scripts" (OuterVolumeSpecName: "scripts") pod "a34cbc14-a8b4-4993-89ba-b36f8a9a4065" (UID: "a34cbc14-a8b4-4993-89ba-b36f8a9a4065"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:33 crc kubenswrapper[4824]: I1209 10:15:33.234007 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a34cbc14-a8b4-4993-89ba-b36f8a9a4065-kube-api-access-zfk6s" (OuterVolumeSpecName: "kube-api-access-zfk6s") pod "a34cbc14-a8b4-4993-89ba-b36f8a9a4065" (UID: "a34cbc14-a8b4-4993-89ba-b36f8a9a4065"). InnerVolumeSpecName "kube-api-access-zfk6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:15:33 crc kubenswrapper[4824]: I1209 10:15:33.261053 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a34cbc14-a8b4-4993-89ba-b36f8a9a4065-config-data" (OuterVolumeSpecName: "config-data") pod "a34cbc14-a8b4-4993-89ba-b36f8a9a4065" (UID: "a34cbc14-a8b4-4993-89ba-b36f8a9a4065"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:33 crc kubenswrapper[4824]: I1209 10:15:33.279107 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a34cbc14-a8b4-4993-89ba-b36f8a9a4065-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a34cbc14-a8b4-4993-89ba-b36f8a9a4065" (UID: "a34cbc14-a8b4-4993-89ba-b36f8a9a4065"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:33 crc kubenswrapper[4824]: I1209 10:15:33.284515 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfk6s\" (UniqueName: \"kubernetes.io/projected/a34cbc14-a8b4-4993-89ba-b36f8a9a4065-kube-api-access-zfk6s\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:33 crc kubenswrapper[4824]: I1209 10:15:33.284569 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a34cbc14-a8b4-4993-89ba-b36f8a9a4065-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:33 crc kubenswrapper[4824]: I1209 10:15:33.284578 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a34cbc14-a8b4-4993-89ba-b36f8a9a4065-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:33 crc kubenswrapper[4824]: I1209 10:15:33.284587 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a34cbc14-a8b4-4993-89ba-b36f8a9a4065-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:33 crc kubenswrapper[4824]: I1209 10:15:33.685072 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-282jv" Dec 09 10:15:33 crc kubenswrapper[4824]: I1209 10:15:33.688454 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-282jv" event={"ID":"a34cbc14-a8b4-4993-89ba-b36f8a9a4065","Type":"ContainerDied","Data":"f7e64ca96567bb5d056db62dd887e885a3cb90a04ca0609a4db893c500c19962"} Dec 09 10:15:33 crc kubenswrapper[4824]: I1209 10:15:33.688513 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7e64ca96567bb5d056db62dd887e885a3cb90a04ca0609a4db893c500c19962" Dec 09 10:15:33 crc kubenswrapper[4824]: I1209 10:15:33.699632 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" exitCode=0 Dec 09 10:15:33 crc kubenswrapper[4824]: I1209 10:15:33.699680 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerDied","Data":"7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1"} Dec 09 10:15:33 crc kubenswrapper[4824]: I1209 10:15:33.699718 4824 scope.go:117] "RemoveContainer" containerID="3844f695598277b223637109f124ad02a65103550c5696a49ebd000720021e3c" Dec 09 10:15:33 crc kubenswrapper[4824]: I1209 10:15:33.701152 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:15:33 crc kubenswrapper[4824]: E1209 10:15:33.701851 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:15:33 crc kubenswrapper[4824]: I1209 10:15:33.984928 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 10:15:33 crc kubenswrapper[4824]: I1209 10:15:33.985427 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="21c56d0b-49a8-4fdd-964f-93a14d1dd003" containerName="nova-api-log" containerID="cri-o://1b314689dbbbace6f1ed19df9e216a45cfac2b1150ebe53f83fc407cff1e7f96" gracePeriod=30 Dec 09 10:15:33 crc kubenswrapper[4824]: I1209 10:15:33.985988 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="21c56d0b-49a8-4fdd-964f-93a14d1dd003" containerName="nova-api-api" containerID="cri-o://32dfc962e7123d7f0622108b65ca08b023a7eafbc684703a1b52ed512f92f127" gracePeriod=30 Dec 09 10:15:34 crc kubenswrapper[4824]: I1209 10:15:34.489506 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 10:15:34 crc kubenswrapper[4824]: I1209 10:15:34.489862 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c" containerName="nova-metadata-log" containerID="cri-o://3dca08d942566f4202f7339b1ab8c1a09137689ed513a884194c5c0b7c8c7b3c" gracePeriod=30 Dec 09 10:15:34 crc kubenswrapper[4824]: I1209 10:15:34.489910 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c" containerName="nova-metadata-metadata" containerID="cri-o://8b9908e6657ec4e808f35ce0be343b0d2e3a72bfb3fb7f6d9ec725e55361459b" gracePeriod=30 Dec 09 10:15:34 crc kubenswrapper[4824]: I1209 10:15:34.506962 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 10:15:34 crc kubenswrapper[4824]: I1209 10:15:34.507207 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="889dfa64-048d-4d55-a734-5b179dfcc1f5" containerName="nova-scheduler-scheduler" containerID="cri-o://12c7d4f6b5c0ed53bdebb01145b627b1460af30b6bbad2f9bd5fc324da920a1f" gracePeriod=30 Dec 09 10:15:34 crc kubenswrapper[4824]: I1209 10:15:34.740560 4824 generic.go:334] "Generic (PLEG): container finished" podID="4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c" containerID="3dca08d942566f4202f7339b1ab8c1a09137689ed513a884194c5c0b7c8c7b3c" exitCode=143 Dec 09 10:15:34 crc kubenswrapper[4824]: I1209 10:15:34.740654 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c","Type":"ContainerDied","Data":"3dca08d942566f4202f7339b1ab8c1a09137689ed513a884194c5c0b7c8c7b3c"} Dec 09 10:15:34 crc kubenswrapper[4824]: I1209 10:15:34.743533 4824 generic.go:334] "Generic (PLEG): container finished" podID="21c56d0b-49a8-4fdd-964f-93a14d1dd003" containerID="32dfc962e7123d7f0622108b65ca08b023a7eafbc684703a1b52ed512f92f127" exitCode=0 Dec 09 10:15:34 crc kubenswrapper[4824]: I1209 10:15:34.743562 4824 generic.go:334] "Generic (PLEG): container finished" podID="21c56d0b-49a8-4fdd-964f-93a14d1dd003" containerID="1b314689dbbbace6f1ed19df9e216a45cfac2b1150ebe53f83fc407cff1e7f96" exitCode=143 Dec 09 10:15:34 crc kubenswrapper[4824]: I1209 10:15:34.743585 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"21c56d0b-49a8-4fdd-964f-93a14d1dd003","Type":"ContainerDied","Data":"32dfc962e7123d7f0622108b65ca08b023a7eafbc684703a1b52ed512f92f127"} Dec 09 10:15:34 crc kubenswrapper[4824]: I1209 10:15:34.743616 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"21c56d0b-49a8-4fdd-964f-93a14d1dd003","Type":"ContainerDied","Data":"1b314689dbbbace6f1ed19df9e216a45cfac2b1150ebe53f83fc407cff1e7f96"} Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.291791 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.353133 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21c56d0b-49a8-4fdd-964f-93a14d1dd003-config-data\") pod \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\" (UID: \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\") " Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.353214 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21c56d0b-49a8-4fdd-964f-93a14d1dd003-internal-tls-certs\") pod \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\" (UID: \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\") " Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.353240 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21c56d0b-49a8-4fdd-964f-93a14d1dd003-public-tls-certs\") pod \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\" (UID: \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\") " Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.353353 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhg46\" (UniqueName: \"kubernetes.io/projected/21c56d0b-49a8-4fdd-964f-93a14d1dd003-kube-api-access-rhg46\") pod \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\" (UID: \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\") " Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.353417 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21c56d0b-49a8-4fdd-964f-93a14d1dd003-logs\") pod \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\" (UID: \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\") " Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.353435 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21c56d0b-49a8-4fdd-964f-93a14d1dd003-combined-ca-bundle\") pod \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\" (UID: \"21c56d0b-49a8-4fdd-964f-93a14d1dd003\") " Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.358080 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21c56d0b-49a8-4fdd-964f-93a14d1dd003-logs" (OuterVolumeSpecName: "logs") pod "21c56d0b-49a8-4fdd-964f-93a14d1dd003" (UID: "21c56d0b-49a8-4fdd-964f-93a14d1dd003"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.362844 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21c56d0b-49a8-4fdd-964f-93a14d1dd003-kube-api-access-rhg46" (OuterVolumeSpecName: "kube-api-access-rhg46") pod "21c56d0b-49a8-4fdd-964f-93a14d1dd003" (UID: "21c56d0b-49a8-4fdd-964f-93a14d1dd003"). InnerVolumeSpecName "kube-api-access-rhg46". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.407792 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21c56d0b-49a8-4fdd-964f-93a14d1dd003-config-data" (OuterVolumeSpecName: "config-data") pod "21c56d0b-49a8-4fdd-964f-93a14d1dd003" (UID: "21c56d0b-49a8-4fdd-964f-93a14d1dd003"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.424760 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21c56d0b-49a8-4fdd-964f-93a14d1dd003-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "21c56d0b-49a8-4fdd-964f-93a14d1dd003" (UID: "21c56d0b-49a8-4fdd-964f-93a14d1dd003"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.443045 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21c56d0b-49a8-4fdd-964f-93a14d1dd003-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "21c56d0b-49a8-4fdd-964f-93a14d1dd003" (UID: "21c56d0b-49a8-4fdd-964f-93a14d1dd003"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.457591 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21c56d0b-49a8-4fdd-964f-93a14d1dd003-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.457642 4824 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21c56d0b-49a8-4fdd-964f-93a14d1dd003-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.457659 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhg46\" (UniqueName: \"kubernetes.io/projected/21c56d0b-49a8-4fdd-964f-93a14d1dd003-kube-api-access-rhg46\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.457675 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21c56d0b-49a8-4fdd-964f-93a14d1dd003-logs\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.457686 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21c56d0b-49a8-4fdd-964f-93a14d1dd003-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.469546 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21c56d0b-49a8-4fdd-964f-93a14d1dd003-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "21c56d0b-49a8-4fdd-964f-93a14d1dd003" (UID: "21c56d0b-49a8-4fdd-964f-93a14d1dd003"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.801156 4824 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21c56d0b-49a8-4fdd-964f-93a14d1dd003-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.821317 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"21c56d0b-49a8-4fdd-964f-93a14d1dd003","Type":"ContainerDied","Data":"c79dcecaa53a8c479cf25b09dab2cbc809af4579142844d46d823400baa3763b"} Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.821420 4824 scope.go:117] "RemoveContainer" containerID="32dfc962e7123d7f0622108b65ca08b023a7eafbc684703a1b52ed512f92f127" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.821836 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.858155 4824 scope.go:117] "RemoveContainer" containerID="1b314689dbbbace6f1ed19df9e216a45cfac2b1150ebe53f83fc407cff1e7f96" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.881343 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.908328 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.939272 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21c56d0b-49a8-4fdd-964f-93a14d1dd003" path="/var/lib/kubelet/pods/21c56d0b-49a8-4fdd-964f-93a14d1dd003/volumes" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.940241 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 09 10:15:35 crc kubenswrapper[4824]: E1209 10:15:35.940754 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6" containerName="init" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.940865 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6" containerName="init" Dec 09 10:15:35 crc kubenswrapper[4824]: E1209 10:15:35.940951 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21c56d0b-49a8-4fdd-964f-93a14d1dd003" containerName="nova-api-log" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.941008 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="21c56d0b-49a8-4fdd-964f-93a14d1dd003" containerName="nova-api-log" Dec 09 10:15:35 crc kubenswrapper[4824]: E1209 10:15:35.941079 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6" containerName="dnsmasq-dns" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.941132 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6" containerName="dnsmasq-dns" Dec 09 10:15:35 crc kubenswrapper[4824]: E1209 10:15:35.941210 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a34cbc14-a8b4-4993-89ba-b36f8a9a4065" containerName="nova-manage" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.941269 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a34cbc14-a8b4-4993-89ba-b36f8a9a4065" containerName="nova-manage" Dec 09 10:15:35 crc kubenswrapper[4824]: E1209 10:15:35.941334 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21c56d0b-49a8-4fdd-964f-93a14d1dd003" containerName="nova-api-api" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.941406 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="21c56d0b-49a8-4fdd-964f-93a14d1dd003" containerName="nova-api-api" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.941712 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a34cbc14-a8b4-4993-89ba-b36f8a9a4065" containerName="nova-manage" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.942072 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="21c56d0b-49a8-4fdd-964f-93a14d1dd003" containerName="nova-api-api" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.942143 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="21c56d0b-49a8-4fdd-964f-93a14d1dd003" containerName="nova-api-log" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.942217 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0049e5d-3e35-4f2f-b303-65d9e8c0a5f6" containerName="dnsmasq-dns" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.943659 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.943945 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.952515 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.952704 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 09 10:15:35 crc kubenswrapper[4824]: I1209 10:15:35.960895 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 09 10:15:36 crc kubenswrapper[4824]: I1209 10:15:36.111276 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fe3dc96-b1f9-475b-9ad7-e752f0cc434e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9fe3dc96-b1f9-475b-9ad7-e752f0cc434e\") " pod="openstack/nova-api-0" Dec 09 10:15:36 crc kubenswrapper[4824]: I1209 10:15:36.111863 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fe3dc96-b1f9-475b-9ad7-e752f0cc434e-public-tls-certs\") pod \"nova-api-0\" (UID: \"9fe3dc96-b1f9-475b-9ad7-e752f0cc434e\") " pod="openstack/nova-api-0" Dec 09 10:15:36 crc kubenswrapper[4824]: I1209 10:15:36.112338 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fe3dc96-b1f9-475b-9ad7-e752f0cc434e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9fe3dc96-b1f9-475b-9ad7-e752f0cc434e\") " pod="openstack/nova-api-0" Dec 09 10:15:36 crc kubenswrapper[4824]: I1209 10:15:36.112633 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9fe3dc96-b1f9-475b-9ad7-e752f0cc434e-logs\") pod \"nova-api-0\" (UID: \"9fe3dc96-b1f9-475b-9ad7-e752f0cc434e\") " pod="openstack/nova-api-0" Dec 09 10:15:36 crc kubenswrapper[4824]: I1209 10:15:36.112831 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln2ct\" (UniqueName: \"kubernetes.io/projected/9fe3dc96-b1f9-475b-9ad7-e752f0cc434e-kube-api-access-ln2ct\") pod \"nova-api-0\" (UID: \"9fe3dc96-b1f9-475b-9ad7-e752f0cc434e\") " pod="openstack/nova-api-0" Dec 09 10:15:36 crc kubenswrapper[4824]: I1209 10:15:36.113112 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fe3dc96-b1f9-475b-9ad7-e752f0cc434e-config-data\") pod \"nova-api-0\" (UID: \"9fe3dc96-b1f9-475b-9ad7-e752f0cc434e\") " pod="openstack/nova-api-0" Dec 09 10:15:36 crc kubenswrapper[4824]: I1209 10:15:36.216202 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fe3dc96-b1f9-475b-9ad7-e752f0cc434e-public-tls-certs\") pod \"nova-api-0\" (UID: \"9fe3dc96-b1f9-475b-9ad7-e752f0cc434e\") " pod="openstack/nova-api-0" Dec 09 10:15:36 crc kubenswrapper[4824]: I1209 10:15:36.216303 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fe3dc96-b1f9-475b-9ad7-e752f0cc434e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9fe3dc96-b1f9-475b-9ad7-e752f0cc434e\") " pod="openstack/nova-api-0" Dec 09 10:15:36 crc kubenswrapper[4824]: I1209 10:15:36.216359 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9fe3dc96-b1f9-475b-9ad7-e752f0cc434e-logs\") pod \"nova-api-0\" (UID: \"9fe3dc96-b1f9-475b-9ad7-e752f0cc434e\") " pod="openstack/nova-api-0" Dec 09 10:15:36 crc kubenswrapper[4824]: I1209 10:15:36.216397 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ln2ct\" (UniqueName: \"kubernetes.io/projected/9fe3dc96-b1f9-475b-9ad7-e752f0cc434e-kube-api-access-ln2ct\") pod \"nova-api-0\" (UID: \"9fe3dc96-b1f9-475b-9ad7-e752f0cc434e\") " pod="openstack/nova-api-0" Dec 09 10:15:36 crc kubenswrapper[4824]: I1209 10:15:36.216443 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fe3dc96-b1f9-475b-9ad7-e752f0cc434e-config-data\") pod \"nova-api-0\" (UID: \"9fe3dc96-b1f9-475b-9ad7-e752f0cc434e\") " pod="openstack/nova-api-0" Dec 09 10:15:36 crc kubenswrapper[4824]: I1209 10:15:36.216487 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fe3dc96-b1f9-475b-9ad7-e752f0cc434e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9fe3dc96-b1f9-475b-9ad7-e752f0cc434e\") " pod="openstack/nova-api-0" Dec 09 10:15:36 crc kubenswrapper[4824]: I1209 10:15:36.217585 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9fe3dc96-b1f9-475b-9ad7-e752f0cc434e-logs\") pod \"nova-api-0\" (UID: \"9fe3dc96-b1f9-475b-9ad7-e752f0cc434e\") " pod="openstack/nova-api-0" Dec 09 10:15:36 crc kubenswrapper[4824]: I1209 10:15:36.221191 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fe3dc96-b1f9-475b-9ad7-e752f0cc434e-public-tls-certs\") pod \"nova-api-0\" (UID: \"9fe3dc96-b1f9-475b-9ad7-e752f0cc434e\") " pod="openstack/nova-api-0" Dec 09 10:15:36 crc kubenswrapper[4824]: I1209 10:15:36.221279 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fe3dc96-b1f9-475b-9ad7-e752f0cc434e-config-data\") pod \"nova-api-0\" (UID: \"9fe3dc96-b1f9-475b-9ad7-e752f0cc434e\") " pod="openstack/nova-api-0" Dec 09 10:15:36 crc kubenswrapper[4824]: I1209 10:15:36.222744 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fe3dc96-b1f9-475b-9ad7-e752f0cc434e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9fe3dc96-b1f9-475b-9ad7-e752f0cc434e\") " pod="openstack/nova-api-0" Dec 09 10:15:36 crc kubenswrapper[4824]: I1209 10:15:36.223431 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fe3dc96-b1f9-475b-9ad7-e752f0cc434e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9fe3dc96-b1f9-475b-9ad7-e752f0cc434e\") " pod="openstack/nova-api-0" Dec 09 10:15:36 crc kubenswrapper[4824]: I1209 10:15:36.244383 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln2ct\" (UniqueName: \"kubernetes.io/projected/9fe3dc96-b1f9-475b-9ad7-e752f0cc434e-kube-api-access-ln2ct\") pod \"nova-api-0\" (UID: \"9fe3dc96-b1f9-475b-9ad7-e752f0cc434e\") " pod="openstack/nova-api-0" Dec 09 10:15:36 crc kubenswrapper[4824]: I1209 10:15:36.283236 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 10:15:36 crc kubenswrapper[4824]: I1209 10:15:36.832980 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 10:15:37 crc kubenswrapper[4824]: E1209 10:15:37.347933 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="12c7d4f6b5c0ed53bdebb01145b627b1460af30b6bbad2f9bd5fc324da920a1f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 10:15:37 crc kubenswrapper[4824]: E1209 10:15:37.352112 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="12c7d4f6b5c0ed53bdebb01145b627b1460af30b6bbad2f9bd5fc324da920a1f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 10:15:37 crc kubenswrapper[4824]: E1209 10:15:37.353605 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="12c7d4f6b5c0ed53bdebb01145b627b1460af30b6bbad2f9bd5fc324da920a1f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 10:15:37 crc kubenswrapper[4824]: E1209 10:15:37.353651 4824 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="889dfa64-048d-4d55-a734-5b179dfcc1f5" containerName="nova-scheduler-scheduler" Dec 09 10:15:37 crc kubenswrapper[4824]: I1209 10:15:37.861315 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9fe3dc96-b1f9-475b-9ad7-e752f0cc434e","Type":"ContainerStarted","Data":"c0e1e5a7abee79731a7a6bd3b3870598bd08a00b35d2542195ee268eb61e3ccb"} Dec 09 10:15:37 crc kubenswrapper[4824]: I1209 10:15:37.861653 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9fe3dc96-b1f9-475b-9ad7-e752f0cc434e","Type":"ContainerStarted","Data":"054a8debd07e9eee0314be21773adf10283eeeee7c40a7f72ed93604bc519a39"} Dec 09 10:15:37 crc kubenswrapper[4824]: I1209 10:15:37.861670 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9fe3dc96-b1f9-475b-9ad7-e752f0cc434e","Type":"ContainerStarted","Data":"396bf7e5ba44a1606b1bb1721040c835fe10610034dcc63fafba115f218b47d3"} Dec 09 10:15:37 crc kubenswrapper[4824]: I1209 10:15:37.865072 4824 generic.go:334] "Generic (PLEG): container finished" podID="384b8d7f-e60a-4421-855d-6d397385f7a8" containerID="b96320be5c824b5fba1b470169895191d2b61ec0a5840ffe6704738b2b7bd816" exitCode=137 Dec 09 10:15:37 crc kubenswrapper[4824]: I1209 10:15:37.865149 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"384b8d7f-e60a-4421-855d-6d397385f7a8","Type":"ContainerDied","Data":"b96320be5c824b5fba1b470169895191d2b61ec0a5840ffe6704738b2b7bd816"} Dec 09 10:15:37 crc kubenswrapper[4824]: I1209 10:15:37.865204 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"384b8d7f-e60a-4421-855d-6d397385f7a8","Type":"ContainerDied","Data":"810012b40b9b0defe7867eaabe35998fa037312da7b903be27c23283f1b795ca"} Dec 09 10:15:37 crc kubenswrapper[4824]: I1209 10:15:37.865214 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="810012b40b9b0defe7867eaabe35998fa037312da7b903be27c23283f1b795ca" Dec 09 10:15:37 crc kubenswrapper[4824]: I1209 10:15:37.879709 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.879685717 podStartE2EDuration="2.879685717s" podCreationTimestamp="2025-12-09 10:15:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:15:37.878380466 +0000 UTC m=+1694.212885133" watchObservedRunningTime="2025-12-09 10:15:37.879685717 +0000 UTC m=+1694.214190384" Dec 09 10:15:37 crc kubenswrapper[4824]: W1209 10:15:37.903854 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21c56d0b_49a8_4fdd_964f_93a14d1dd003.slice/crio-1b314689dbbbace6f1ed19df9e216a45cfac2b1150ebe53f83fc407cff1e7f96.scope WatchSource:0}: Error finding container 1b314689dbbbace6f1ed19df9e216a45cfac2b1150ebe53f83fc407cff1e7f96: Status 404 returned error can't find the container with id 1b314689dbbbace6f1ed19df9e216a45cfac2b1150ebe53f83fc407cff1e7f96 Dec 09 10:15:37 crc kubenswrapper[4824]: W1209 10:15:37.905140 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21c56d0b_49a8_4fdd_964f_93a14d1dd003.slice/crio-32dfc962e7123d7f0622108b65ca08b023a7eafbc684703a1b52ed512f92f127.scope WatchSource:0}: Error finding container 32dfc962e7123d7f0622108b65ca08b023a7eafbc684703a1b52ed512f92f127: Status 404 returned error can't find the container with id 32dfc962e7123d7f0622108b65ca08b023a7eafbc684703a1b52ed512f92f127 Dec 09 10:15:38 crc kubenswrapper[4824]: I1209 10:15:38.181743 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 09 10:15:38 crc kubenswrapper[4824]: I1209 10:15:38.278985 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/384b8d7f-e60a-4421-855d-6d397385f7a8-combined-ca-bundle\") pod \"384b8d7f-e60a-4421-855d-6d397385f7a8\" (UID: \"384b8d7f-e60a-4421-855d-6d397385f7a8\") " Dec 09 10:15:38 crc kubenswrapper[4824]: I1209 10:15:38.279368 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/384b8d7f-e60a-4421-855d-6d397385f7a8-scripts\") pod \"384b8d7f-e60a-4421-855d-6d397385f7a8\" (UID: \"384b8d7f-e60a-4421-855d-6d397385f7a8\") " Dec 09 10:15:38 crc kubenswrapper[4824]: I1209 10:15:38.279637 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/384b8d7f-e60a-4421-855d-6d397385f7a8-config-data\") pod \"384b8d7f-e60a-4421-855d-6d397385f7a8\" (UID: \"384b8d7f-e60a-4421-855d-6d397385f7a8\") " Dec 09 10:15:38 crc kubenswrapper[4824]: I1209 10:15:38.279673 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbjwc\" (UniqueName: \"kubernetes.io/projected/384b8d7f-e60a-4421-855d-6d397385f7a8-kube-api-access-qbjwc\") pod \"384b8d7f-e60a-4421-855d-6d397385f7a8\" (UID: \"384b8d7f-e60a-4421-855d-6d397385f7a8\") " Dec 09 10:15:38 crc kubenswrapper[4824]: I1209 10:15:38.285998 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/384b8d7f-e60a-4421-855d-6d397385f7a8-scripts" (OuterVolumeSpecName: "scripts") pod "384b8d7f-e60a-4421-855d-6d397385f7a8" (UID: "384b8d7f-e60a-4421-855d-6d397385f7a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:38 crc kubenswrapper[4824]: I1209 10:15:38.286055 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/384b8d7f-e60a-4421-855d-6d397385f7a8-kube-api-access-qbjwc" (OuterVolumeSpecName: "kube-api-access-qbjwc") pod "384b8d7f-e60a-4421-855d-6d397385f7a8" (UID: "384b8d7f-e60a-4421-855d-6d397385f7a8"). InnerVolumeSpecName "kube-api-access-qbjwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:15:38 crc kubenswrapper[4824]: E1209 10:15:38.332963 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b1dd9ff_c6e7_4b2f_92c2_b0c020b5783c.slice/crio-conmon-8b9908e6657ec4e808f35ce0be343b0d2e3a72bfb3fb7f6d9ec725e55361459b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b1dd9ff_c6e7_4b2f_92c2_b0c020b5783c.slice/crio-8b9908e6657ec4e808f35ce0be343b0d2e3a72bfb3fb7f6d9ec725e55361459b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod384b8d7f_e60a_4421_855d_6d397385f7a8.slice/crio-b96320be5c824b5fba1b470169895191d2b61ec0a5840ffe6704738b2b7bd816.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod384b8d7f_e60a_4421_855d_6d397385f7a8.slice/crio-conmon-b96320be5c824b5fba1b470169895191d2b61ec0a5840ffe6704738b2b7bd816.scope\": RecentStats: unable to find data in memory cache]" Dec 09 10:15:38 crc kubenswrapper[4824]: I1209 10:15:38.383245 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbjwc\" (UniqueName: \"kubernetes.io/projected/384b8d7f-e60a-4421-855d-6d397385f7a8-kube-api-access-qbjwc\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:38 crc kubenswrapper[4824]: I1209 10:15:38.383300 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/384b8d7f-e60a-4421-855d-6d397385f7a8-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:38 crc kubenswrapper[4824]: I1209 10:15:38.438107 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/384b8d7f-e60a-4421-855d-6d397385f7a8-config-data" (OuterVolumeSpecName: "config-data") pod "384b8d7f-e60a-4421-855d-6d397385f7a8" (UID: "384b8d7f-e60a-4421-855d-6d397385f7a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:38 crc kubenswrapper[4824]: I1209 10:15:38.478125 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/384b8d7f-e60a-4421-855d-6d397385f7a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "384b8d7f-e60a-4421-855d-6d397385f7a8" (UID: "384b8d7f-e60a-4421-855d-6d397385f7a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:38 crc kubenswrapper[4824]: I1209 10:15:38.498623 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/384b8d7f-e60a-4421-855d-6d397385f7a8-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:38 crc kubenswrapper[4824]: I1209 10:15:38.498659 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/384b8d7f-e60a-4421-855d-6d397385f7a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.044277 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.120917 4824 generic.go:334] "Generic (PLEG): container finished" podID="4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c" containerID="8b9908e6657ec4e808f35ce0be343b0d2e3a72bfb3fb7f6d9ec725e55361459b" exitCode=0 Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.121049 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.121850 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.122727 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c","Type":"ContainerDied","Data":"8b9908e6657ec4e808f35ce0be343b0d2e3a72bfb3fb7f6d9ec725e55361459b"} Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.122769 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c","Type":"ContainerDied","Data":"8ac6f318e9cd03ac199b5d9a39189b8f2123d11c787b916e6485dd7a3e72e8cf"} Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.122796 4824 scope.go:117] "RemoveContainer" containerID="8b9908e6657ec4e808f35ce0be343b0d2e3a72bfb3fb7f6d9ec725e55361459b" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.249022 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-logs\") pod \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\" (UID: \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\") " Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.249215 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-config-data\") pod \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\" (UID: \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\") " Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.249256 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-nova-metadata-tls-certs\") pod \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\" (UID: \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\") " Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.249277 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-combined-ca-bundle\") pod \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\" (UID: \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\") " Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.249343 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4rwc\" (UniqueName: \"kubernetes.io/projected/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-kube-api-access-m4rwc\") pod \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\" (UID: \"4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c\") " Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.252279 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-logs" (OuterVolumeSpecName: "logs") pod "4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c" (UID: "4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.261189 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.262448 4824 scope.go:117] "RemoveContainer" containerID="3dca08d942566f4202f7339b1ab8c1a09137689ed513a884194c5c0b7c8c7b3c" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.276032 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.292470 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-config-data" (OuterVolumeSpecName: "config-data") pod "4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c" (UID: "4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.300611 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-kube-api-access-m4rwc" (OuterVolumeSpecName: "kube-api-access-m4rwc") pod "4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c" (UID: "4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c"). InnerVolumeSpecName "kube-api-access-m4rwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.315639 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 09 10:15:39 crc kubenswrapper[4824]: E1209 10:15:39.325575 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="384b8d7f-e60a-4421-855d-6d397385f7a8" containerName="aodh-api" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.325625 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="384b8d7f-e60a-4421-855d-6d397385f7a8" containerName="aodh-api" Dec 09 10:15:39 crc kubenswrapper[4824]: E1209 10:15:39.325650 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="384b8d7f-e60a-4421-855d-6d397385f7a8" containerName="aodh-evaluator" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.325658 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="384b8d7f-e60a-4421-855d-6d397385f7a8" containerName="aodh-evaluator" Dec 09 10:15:39 crc kubenswrapper[4824]: E1209 10:15:39.325694 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="384b8d7f-e60a-4421-855d-6d397385f7a8" containerName="aodh-notifier" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.325702 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="384b8d7f-e60a-4421-855d-6d397385f7a8" containerName="aodh-notifier" Dec 09 10:15:39 crc kubenswrapper[4824]: E1209 10:15:39.325717 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="384b8d7f-e60a-4421-855d-6d397385f7a8" containerName="aodh-listener" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.325725 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="384b8d7f-e60a-4421-855d-6d397385f7a8" containerName="aodh-listener" Dec 09 10:15:39 crc kubenswrapper[4824]: E1209 10:15:39.325800 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c" containerName="nova-metadata-log" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.325810 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c" containerName="nova-metadata-log" Dec 09 10:15:39 crc kubenswrapper[4824]: E1209 10:15:39.325830 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c" containerName="nova-metadata-metadata" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.325837 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c" containerName="nova-metadata-metadata" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.326130 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="384b8d7f-e60a-4421-855d-6d397385f7a8" containerName="aodh-listener" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.326165 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c" containerName="nova-metadata-log" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.326193 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="384b8d7f-e60a-4421-855d-6d397385f7a8" containerName="aodh-notifier" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.326207 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="384b8d7f-e60a-4421-855d-6d397385f7a8" containerName="aodh-evaluator" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.326220 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="384b8d7f-e60a-4421-855d-6d397385f7a8" containerName="aodh-api" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.326244 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c" containerName="nova-metadata-metadata" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.329047 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.332548 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.332746 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.334753 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.335102 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-mgzrr" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.335365 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.337318 4824 scope.go:117] "RemoveContainer" containerID="8b9908e6657ec4e808f35ce0be343b0d2e3a72bfb3fb7f6d9ec725e55361459b" Dec 09 10:15:39 crc kubenswrapper[4824]: E1209 10:15:39.342141 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b9908e6657ec4e808f35ce0be343b0d2e3a72bfb3fb7f6d9ec725e55361459b\": container with ID starting with 8b9908e6657ec4e808f35ce0be343b0d2e3a72bfb3fb7f6d9ec725e55361459b not found: ID does not exist" containerID="8b9908e6657ec4e808f35ce0be343b0d2e3a72bfb3fb7f6d9ec725e55361459b" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.342181 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b9908e6657ec4e808f35ce0be343b0d2e3a72bfb3fb7f6d9ec725e55361459b"} err="failed to get container status \"8b9908e6657ec4e808f35ce0be343b0d2e3a72bfb3fb7f6d9ec725e55361459b\": rpc error: code = NotFound desc = could not find container \"8b9908e6657ec4e808f35ce0be343b0d2e3a72bfb3fb7f6d9ec725e55361459b\": container with ID starting with 8b9908e6657ec4e808f35ce0be343b0d2e3a72bfb3fb7f6d9ec725e55361459b not found: ID does not exist" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.342209 4824 scope.go:117] "RemoveContainer" containerID="3dca08d942566f4202f7339b1ab8c1a09137689ed513a884194c5c0b7c8c7b3c" Dec 09 10:15:39 crc kubenswrapper[4824]: E1209 10:15:39.342942 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dca08d942566f4202f7339b1ab8c1a09137689ed513a884194c5c0b7c8c7b3c\": container with ID starting with 3dca08d942566f4202f7339b1ab8c1a09137689ed513a884194c5c0b7c8c7b3c not found: ID does not exist" containerID="3dca08d942566f4202f7339b1ab8c1a09137689ed513a884194c5c0b7c8c7b3c" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.342963 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dca08d942566f4202f7339b1ab8c1a09137689ed513a884194c5c0b7c8c7b3c"} err="failed to get container status \"3dca08d942566f4202f7339b1ab8c1a09137689ed513a884194c5c0b7c8c7b3c\": rpc error: code = NotFound desc = could not find container \"3dca08d942566f4202f7339b1ab8c1a09137689ed513a884194c5c0b7c8c7b3c\": container with ID starting with 3dca08d942566f4202f7339b1ab8c1a09137689ed513a884194c5c0b7c8c7b3c not found: ID does not exist" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.354913 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\") " pod="openstack/aodh-0" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.354987 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-internal-tls-certs\") pod \"aodh-0\" (UID: \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\") " pod="openstack/aodh-0" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.355046 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-public-tls-certs\") pod \"aodh-0\" (UID: \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\") " pod="openstack/aodh-0" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.355063 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-config-data\") pod \"aodh-0\" (UID: \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\") " pod="openstack/aodh-0" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.355108 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-scripts\") pod \"aodh-0\" (UID: \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\") " pod="openstack/aodh-0" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.355191 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbxp8\" (UniqueName: \"kubernetes.io/projected/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-kube-api-access-mbxp8\") pod \"aodh-0\" (UID: \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\") " pod="openstack/aodh-0" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.355304 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4rwc\" (UniqueName: \"kubernetes.io/projected/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-kube-api-access-m4rwc\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.355315 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-logs\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.355323 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.364242 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.404921 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c" (UID: "4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.425941 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c" (UID: "4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.457001 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-scripts\") pod \"aodh-0\" (UID: \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\") " pod="openstack/aodh-0" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.457132 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbxp8\" (UniqueName: \"kubernetes.io/projected/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-kube-api-access-mbxp8\") pod \"aodh-0\" (UID: \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\") " pod="openstack/aodh-0" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.457255 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\") " pod="openstack/aodh-0" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.457301 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-internal-tls-certs\") pod \"aodh-0\" (UID: \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\") " pod="openstack/aodh-0" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.457358 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-public-tls-certs\") pod \"aodh-0\" (UID: \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\") " pod="openstack/aodh-0" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.457375 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-config-data\") pod \"aodh-0\" (UID: \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\") " pod="openstack/aodh-0" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.457430 4824 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.457441 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.461431 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-config-data\") pod \"aodh-0\" (UID: \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\") " pod="openstack/aodh-0" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.461761 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-scripts\") pod \"aodh-0\" (UID: \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\") " pod="openstack/aodh-0" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.464919 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\") " pod="openstack/aodh-0" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.465446 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-internal-tls-certs\") pod \"aodh-0\" (UID: \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\") " pod="openstack/aodh-0" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.466207 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-public-tls-certs\") pod \"aodh-0\" (UID: \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\") " pod="openstack/aodh-0" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.478277 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbxp8\" (UniqueName: \"kubernetes.io/projected/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-kube-api-access-mbxp8\") pod \"aodh-0\" (UID: \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\") " pod="openstack/aodh-0" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.704635 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.758704 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 10:15:39 crc kubenswrapper[4824]: I1209 10:15:39.785683 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.101246 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="384b8d7f-e60a-4421-855d-6d397385f7a8" path="/var/lib/kubelet/pods/384b8d7f-e60a-4421-855d-6d397385f7a8/volumes" Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.104912 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c" path="/var/lib/kubelet/pods/4b1dd9ff-c6e7-4b2f-92c2-b0c020b5783c/volumes" Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.110180 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.112716 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.112895 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.120263 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.120607 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.251147 4824 generic.go:334] "Generic (PLEG): container finished" podID="889dfa64-048d-4d55-a734-5b179dfcc1f5" containerID="12c7d4f6b5c0ed53bdebb01145b627b1460af30b6bbad2f9bd5fc324da920a1f" exitCode=0 Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.251200 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"889dfa64-048d-4d55-a734-5b179dfcc1f5","Type":"ContainerDied","Data":"12c7d4f6b5c0ed53bdebb01145b627b1460af30b6bbad2f9bd5fc324da920a1f"} Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.284130 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbnjc\" (UniqueName: \"kubernetes.io/projected/b295ca7f-46ac-4ab1-afc4-f1554b84c0d3-kube-api-access-cbnjc\") pod \"nova-metadata-0\" (UID: \"b295ca7f-46ac-4ab1-afc4-f1554b84c0d3\") " pod="openstack/nova-metadata-0" Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.284289 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b295ca7f-46ac-4ab1-afc4-f1554b84c0d3-config-data\") pod \"nova-metadata-0\" (UID: \"b295ca7f-46ac-4ab1-afc4-f1554b84c0d3\") " pod="openstack/nova-metadata-0" Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.284437 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b295ca7f-46ac-4ab1-afc4-f1554b84c0d3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b295ca7f-46ac-4ab1-afc4-f1554b84c0d3\") " pod="openstack/nova-metadata-0" Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.284491 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b295ca7f-46ac-4ab1-afc4-f1554b84c0d3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b295ca7f-46ac-4ab1-afc4-f1554b84c0d3\") " pod="openstack/nova-metadata-0" Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.284618 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b295ca7f-46ac-4ab1-afc4-f1554b84c0d3-logs\") pod \"nova-metadata-0\" (UID: \"b295ca7f-46ac-4ab1-afc4-f1554b84c0d3\") " pod="openstack/nova-metadata-0" Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.539573 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b295ca7f-46ac-4ab1-afc4-f1554b84c0d3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b295ca7f-46ac-4ab1-afc4-f1554b84c0d3\") " pod="openstack/nova-metadata-0" Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.540048 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b295ca7f-46ac-4ab1-afc4-f1554b84c0d3-logs\") pod \"nova-metadata-0\" (UID: \"b295ca7f-46ac-4ab1-afc4-f1554b84c0d3\") " pod="openstack/nova-metadata-0" Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.540276 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbnjc\" (UniqueName: \"kubernetes.io/projected/b295ca7f-46ac-4ab1-afc4-f1554b84c0d3-kube-api-access-cbnjc\") pod \"nova-metadata-0\" (UID: \"b295ca7f-46ac-4ab1-afc4-f1554b84c0d3\") " pod="openstack/nova-metadata-0" Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.540437 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b295ca7f-46ac-4ab1-afc4-f1554b84c0d3-config-data\") pod \"nova-metadata-0\" (UID: \"b295ca7f-46ac-4ab1-afc4-f1554b84c0d3\") " pod="openstack/nova-metadata-0" Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.540624 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b295ca7f-46ac-4ab1-afc4-f1554b84c0d3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b295ca7f-46ac-4ab1-afc4-f1554b84c0d3\") " pod="openstack/nova-metadata-0" Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.542196 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b295ca7f-46ac-4ab1-afc4-f1554b84c0d3-logs\") pod \"nova-metadata-0\" (UID: \"b295ca7f-46ac-4ab1-afc4-f1554b84c0d3\") " pod="openstack/nova-metadata-0" Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.593466 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbnjc\" (UniqueName: \"kubernetes.io/projected/b295ca7f-46ac-4ab1-afc4-f1554b84c0d3-kube-api-access-cbnjc\") pod \"nova-metadata-0\" (UID: \"b295ca7f-46ac-4ab1-afc4-f1554b84c0d3\") " pod="openstack/nova-metadata-0" Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.603532 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b295ca7f-46ac-4ab1-afc4-f1554b84c0d3-config-data\") pod \"nova-metadata-0\" (UID: \"b295ca7f-46ac-4ab1-afc4-f1554b84c0d3\") " pod="openstack/nova-metadata-0" Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.610856 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b295ca7f-46ac-4ab1-afc4-f1554b84c0d3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b295ca7f-46ac-4ab1-afc4-f1554b84c0d3\") " pod="openstack/nova-metadata-0" Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.614955 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b295ca7f-46ac-4ab1-afc4-f1554b84c0d3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b295ca7f-46ac-4ab1-afc4-f1554b84c0d3\") " pod="openstack/nova-metadata-0" Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.844047 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 10:15:40 crc kubenswrapper[4824]: I1209 10:15:40.951356 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 09 10:15:40 crc kubenswrapper[4824]: W1209 10:15:40.964715 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9d7c93f_e88e_4b55_91b8_f9dc9f42b912.slice/crio-8d712badb43050a0c7c9b75963f99c41b35c44edc9ac1c5bd070f05ad7b5dc6c WatchSource:0}: Error finding container 8d712badb43050a0c7c9b75963f99c41b35c44edc9ac1c5bd070f05ad7b5dc6c: Status 404 returned error can't find the container with id 8d712badb43050a0c7c9b75963f99c41b35c44edc9ac1c5bd070f05ad7b5dc6c Dec 09 10:15:41 crc kubenswrapper[4824]: I1209 10:15:41.116867 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 10:15:41 crc kubenswrapper[4824]: I1209 10:15:41.406117 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/889dfa64-048d-4d55-a734-5b179dfcc1f5-config-data\") pod \"889dfa64-048d-4d55-a734-5b179dfcc1f5\" (UID: \"889dfa64-048d-4d55-a734-5b179dfcc1f5\") " Dec 09 10:15:41 crc kubenswrapper[4824]: I1209 10:15:41.406365 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m69tb\" (UniqueName: \"kubernetes.io/projected/889dfa64-048d-4d55-a734-5b179dfcc1f5-kube-api-access-m69tb\") pod \"889dfa64-048d-4d55-a734-5b179dfcc1f5\" (UID: \"889dfa64-048d-4d55-a734-5b179dfcc1f5\") " Dec 09 10:15:41 crc kubenswrapper[4824]: I1209 10:15:41.406662 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/889dfa64-048d-4d55-a734-5b179dfcc1f5-combined-ca-bundle\") pod \"889dfa64-048d-4d55-a734-5b179dfcc1f5\" (UID: \"889dfa64-048d-4d55-a734-5b179dfcc1f5\") " Dec 09 10:15:41 crc kubenswrapper[4824]: I1209 10:15:41.444540 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/889dfa64-048d-4d55-a734-5b179dfcc1f5-kube-api-access-m69tb" (OuterVolumeSpecName: "kube-api-access-m69tb") pod "889dfa64-048d-4d55-a734-5b179dfcc1f5" (UID: "889dfa64-048d-4d55-a734-5b179dfcc1f5"). InnerVolumeSpecName "kube-api-access-m69tb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:15:41 crc kubenswrapper[4824]: I1209 10:15:41.469333 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912","Type":"ContainerStarted","Data":"8d712badb43050a0c7c9b75963f99c41b35c44edc9ac1c5bd070f05ad7b5dc6c"} Dec 09 10:15:41 crc kubenswrapper[4824]: I1209 10:15:41.508134 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"889dfa64-048d-4d55-a734-5b179dfcc1f5","Type":"ContainerDied","Data":"bfc705c8e1c2213ec386358afc1fde6e001bd5c5d5abeaaec73e13053dc09915"} Dec 09 10:15:41 crc kubenswrapper[4824]: I1209 10:15:41.508184 4824 scope.go:117] "RemoveContainer" containerID="12c7d4f6b5c0ed53bdebb01145b627b1460af30b6bbad2f9bd5fc324da920a1f" Dec 09 10:15:41 crc kubenswrapper[4824]: I1209 10:15:41.510302 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 10:15:41 crc kubenswrapper[4824]: I1209 10:15:41.512461 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m69tb\" (UniqueName: \"kubernetes.io/projected/889dfa64-048d-4d55-a734-5b179dfcc1f5-kube-api-access-m69tb\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:41 crc kubenswrapper[4824]: I1209 10:15:41.570691 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/889dfa64-048d-4d55-a734-5b179dfcc1f5-config-data" (OuterVolumeSpecName: "config-data") pod "889dfa64-048d-4d55-a734-5b179dfcc1f5" (UID: "889dfa64-048d-4d55-a734-5b179dfcc1f5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:41 crc kubenswrapper[4824]: I1209 10:15:41.575169 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/889dfa64-048d-4d55-a734-5b179dfcc1f5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "889dfa64-048d-4d55-a734-5b179dfcc1f5" (UID: "889dfa64-048d-4d55-a734-5b179dfcc1f5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:41 crc kubenswrapper[4824]: I1209 10:15:41.614182 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/889dfa64-048d-4d55-a734-5b179dfcc1f5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:41 crc kubenswrapper[4824]: I1209 10:15:41.614425 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/889dfa64-048d-4d55-a734-5b179dfcc1f5-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:42 crc kubenswrapper[4824]: I1209 10:15:42.027149 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 10:15:42 crc kubenswrapper[4824]: I1209 10:15:42.053663 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 10:15:42 crc kubenswrapper[4824]: I1209 10:15:42.086806 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 10:15:42 crc kubenswrapper[4824]: I1209 10:15:42.110949 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 10:15:42 crc kubenswrapper[4824]: E1209 10:15:42.111696 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="889dfa64-048d-4d55-a734-5b179dfcc1f5" containerName="nova-scheduler-scheduler" Dec 09 10:15:42 crc kubenswrapper[4824]: I1209 10:15:42.111724 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="889dfa64-048d-4d55-a734-5b179dfcc1f5" containerName="nova-scheduler-scheduler" Dec 09 10:15:42 crc kubenswrapper[4824]: I1209 10:15:42.112053 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="889dfa64-048d-4d55-a734-5b179dfcc1f5" containerName="nova-scheduler-scheduler" Dec 09 10:15:42 crc kubenswrapper[4824]: I1209 10:15:42.113093 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 10:15:42 crc kubenswrapper[4824]: I1209 10:15:42.119531 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 09 10:15:42 crc kubenswrapper[4824]: I1209 10:15:42.122540 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9hfw\" (UniqueName: \"kubernetes.io/projected/543b3a99-2d5c-4467-8076-b284ea519244-kube-api-access-p9hfw\") pod \"nova-scheduler-0\" (UID: \"543b3a99-2d5c-4467-8076-b284ea519244\") " pod="openstack/nova-scheduler-0" Dec 09 10:15:42 crc kubenswrapper[4824]: I1209 10:15:42.122750 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/543b3a99-2d5c-4467-8076-b284ea519244-config-data\") pod \"nova-scheduler-0\" (UID: \"543b3a99-2d5c-4467-8076-b284ea519244\") " pod="openstack/nova-scheduler-0" Dec 09 10:15:42 crc kubenswrapper[4824]: I1209 10:15:42.122831 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/543b3a99-2d5c-4467-8076-b284ea519244-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"543b3a99-2d5c-4467-8076-b284ea519244\") " pod="openstack/nova-scheduler-0" Dec 09 10:15:42 crc kubenswrapper[4824]: I1209 10:15:42.127347 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 10:15:42 crc kubenswrapper[4824]: I1209 10:15:42.223840 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/543b3a99-2d5c-4467-8076-b284ea519244-config-data\") pod \"nova-scheduler-0\" (UID: \"543b3a99-2d5c-4467-8076-b284ea519244\") " pod="openstack/nova-scheduler-0" Dec 09 10:15:42 crc kubenswrapper[4824]: I1209 10:15:42.223916 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/543b3a99-2d5c-4467-8076-b284ea519244-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"543b3a99-2d5c-4467-8076-b284ea519244\") " pod="openstack/nova-scheduler-0" Dec 09 10:15:42 crc kubenswrapper[4824]: I1209 10:15:42.224042 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9hfw\" (UniqueName: \"kubernetes.io/projected/543b3a99-2d5c-4467-8076-b284ea519244-kube-api-access-p9hfw\") pod \"nova-scheduler-0\" (UID: \"543b3a99-2d5c-4467-8076-b284ea519244\") " pod="openstack/nova-scheduler-0" Dec 09 10:15:42 crc kubenswrapper[4824]: I1209 10:15:42.228981 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/543b3a99-2d5c-4467-8076-b284ea519244-config-data\") pod \"nova-scheduler-0\" (UID: \"543b3a99-2d5c-4467-8076-b284ea519244\") " pod="openstack/nova-scheduler-0" Dec 09 10:15:42 crc kubenswrapper[4824]: I1209 10:15:42.230215 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/543b3a99-2d5c-4467-8076-b284ea519244-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"543b3a99-2d5c-4467-8076-b284ea519244\") " pod="openstack/nova-scheduler-0" Dec 09 10:15:42 crc kubenswrapper[4824]: I1209 10:15:42.250425 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9hfw\" (UniqueName: \"kubernetes.io/projected/543b3a99-2d5c-4467-8076-b284ea519244-kube-api-access-p9hfw\") pod \"nova-scheduler-0\" (UID: \"543b3a99-2d5c-4467-8076-b284ea519244\") " pod="openstack/nova-scheduler-0" Dec 09 10:15:42 crc kubenswrapper[4824]: I1209 10:15:42.436198 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 10:15:42 crc kubenswrapper[4824]: I1209 10:15:42.795172 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912","Type":"ContainerStarted","Data":"996c1915e4732b450d51c25f2a00141f01516d9c5baedbaa1da35c6a1ffe3be2"} Dec 09 10:15:42 crc kubenswrapper[4824]: I1209 10:15:42.796903 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b295ca7f-46ac-4ab1-afc4-f1554b84c0d3","Type":"ContainerStarted","Data":"e4d94707b121673a427a08396d9c5701729b7ec7cc6c4ba75fafc103104a8d49"} Dec 09 10:15:42 crc kubenswrapper[4824]: I1209 10:15:42.796943 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b295ca7f-46ac-4ab1-afc4-f1554b84c0d3","Type":"ContainerStarted","Data":"862e5ff078ee4e442f07d8b93eb7449ddb93566f2da078cf350a3a78e921a688"} Dec 09 10:15:43 crc kubenswrapper[4824]: I1209 10:15:43.405584 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 10:15:43 crc kubenswrapper[4824]: W1209 10:15:43.406560 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod543b3a99_2d5c_4467_8076_b284ea519244.slice/crio-1c253dc5ac5cdd3bf3c1a8bdd04a41d17b18a84fe9c118479a78c81fd643b931 WatchSource:0}: Error finding container 1c253dc5ac5cdd3bf3c1a8bdd04a41d17b18a84fe9c118479a78c81fd643b931: Status 404 returned error can't find the container with id 1c253dc5ac5cdd3bf3c1a8bdd04a41d17b18a84fe9c118479a78c81fd643b931 Dec 09 10:15:44 crc kubenswrapper[4824]: I1209 10:15:44.259612 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="889dfa64-048d-4d55-a734-5b179dfcc1f5" path="/var/lib/kubelet/pods/889dfa64-048d-4d55-a734-5b179dfcc1f5/volumes" Dec 09 10:15:44 crc kubenswrapper[4824]: I1209 10:15:44.767774 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-vz44w container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 10:15:44 crc kubenswrapper[4824]: I1209 10:15:44.767925 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" podUID="93c989b8-7b86-4339-bbab-5886c7d13dc9" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.55:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 10:15:44 crc kubenswrapper[4824]: I1209 10:15:44.775996 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-vz44w container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:8083/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 10:15:44 crc kubenswrapper[4824]: I1209 10:15:44.776095 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" podUID="93c989b8-7b86-4339-bbab-5886c7d13dc9" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.55:8083/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 10:15:44 crc kubenswrapper[4824]: I1209 10:15:44.800473 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b295ca7f-46ac-4ab1-afc4-f1554b84c0d3","Type":"ContainerStarted","Data":"8cf481f3e0f514fc9ee1f968b4ef3dd05a88cd3d1190b7fbacba38fbbcd085da"} Dec 09 10:15:44 crc kubenswrapper[4824]: I1209 10:15:44.849018 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"543b3a99-2d5c-4467-8076-b284ea519244","Type":"ContainerStarted","Data":"1c253dc5ac5cdd3bf3c1a8bdd04a41d17b18a84fe9c118479a78c81fd643b931"} Dec 09 10:15:44 crc kubenswrapper[4824]: I1209 10:15:44.875599 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912","Type":"ContainerStarted","Data":"990c7f6cffbc366e0e2c8772c94ef0716443d4bfc59f04eb74f5fd264c37e9a2"} Dec 09 10:15:45 crc kubenswrapper[4824]: I1209 10:15:45.030534 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=6.03049646 podStartE2EDuration="6.03049646s" podCreationTimestamp="2025-12-09 10:15:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:15:44.971570458 +0000 UTC m=+1701.306075145" watchObservedRunningTime="2025-12-09 10:15:45.03049646 +0000 UTC m=+1701.365001127" Dec 09 10:15:46 crc kubenswrapper[4824]: I1209 10:15:46.368826 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 09 10:15:46 crc kubenswrapper[4824]: I1209 10:15:46.369326 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 09 10:15:46 crc kubenswrapper[4824]: I1209 10:15:46.369344 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 09 10:15:46 crc kubenswrapper[4824]: I1209 10:15:46.369354 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 09 10:15:46 crc kubenswrapper[4824]: I1209 10:15:46.376081 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912","Type":"ContainerStarted","Data":"92c378f182312eddc1cf9b26ea46a7f2ca06588876fda5a793d6e23092f4d12a"} Dec 09 10:15:46 crc kubenswrapper[4824]: I1209 10:15:46.398389 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"543b3a99-2d5c-4467-8076-b284ea519244","Type":"ContainerStarted","Data":"3b7641eb20cb56e9ff9c2498c4508363797dfcc34b95c3ed748088fb64ca821d"} Dec 09 10:15:46 crc kubenswrapper[4824]: I1209 10:15:46.420569 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=5.420549701 podStartE2EDuration="5.420549701s" podCreationTimestamp="2025-12-09 10:15:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:15:46.415338817 +0000 UTC m=+1702.749843484" watchObservedRunningTime="2025-12-09 10:15:46.420549701 +0000 UTC m=+1702.755054368" Dec 09 10:15:47 crc kubenswrapper[4824]: I1209 10:15:46.997077 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:15:47 crc kubenswrapper[4824]: E1209 10:15:46.997450 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:15:47 crc kubenswrapper[4824]: I1209 10:15:47.656522 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 09 10:15:48 crc kubenswrapper[4824]: I1209 10:15:47.688015 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9fe3dc96-b1f9-475b-9ad7-e752f0cc434e" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.4:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 10:15:48 crc kubenswrapper[4824]: I1209 10:15:47.688439 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9fe3dc96-b1f9-475b-9ad7-e752f0cc434e" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.4:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 10:15:50 crc kubenswrapper[4824]: I1209 10:15:50.100737 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912","Type":"ContainerStarted","Data":"fb69b48a84b960f40b8160e2b122bf60e8d692316bc07251392c08bfa27737b6"} Dec 09 10:15:50 crc kubenswrapper[4824]: I1209 10:15:50.155894 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=5.111205074 podStartE2EDuration="11.155872636s" podCreationTimestamp="2025-12-09 10:15:39 +0000 UTC" firstStartedPulling="2025-12-09 10:15:40.972295074 +0000 UTC m=+1697.306799741" lastFinishedPulling="2025-12-09 10:15:47.016962636 +0000 UTC m=+1703.351467303" observedRunningTime="2025-12-09 10:15:50.146944864 +0000 UTC m=+1706.481449531" watchObservedRunningTime="2025-12-09 10:15:50.155872636 +0000 UTC m=+1706.490377303" Dec 09 10:15:50 crc kubenswrapper[4824]: I1209 10:15:50.547844 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 10:15:50 crc kubenswrapper[4824]: I1209 10:15:50.662555 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="d254951a-1dd1-4b09-8cfa-e765853cb5e9" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 09 10:15:50 crc kubenswrapper[4824]: I1209 10:15:50.847655 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 09 10:15:50 crc kubenswrapper[4824]: I1209 10:15:50.847692 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 09 10:15:51 crc kubenswrapper[4824]: I1209 10:15:51.863180 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b295ca7f-46ac-4ab1-afc4-f1554b84c0d3" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.6:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 10:15:51 crc kubenswrapper[4824]: I1209 10:15:51.863190 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b295ca7f-46ac-4ab1-afc4-f1554b84c0d3" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.6:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 10:15:52 crc kubenswrapper[4824]: I1209 10:15:52.436401 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 09 10:15:52 crc kubenswrapper[4824]: I1209 10:15:52.486633 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 09 10:15:53 crc kubenswrapper[4824]: I1209 10:15:53.174553 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 09 10:15:56 crc kubenswrapper[4824]: I1209 10:15:56.413053 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 09 10:15:56 crc kubenswrapper[4824]: I1209 10:15:56.414100 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 09 10:15:56 crc kubenswrapper[4824]: I1209 10:15:56.414386 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 09 10:15:56 crc kubenswrapper[4824]: I1209 10:15:56.420798 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 09 10:15:56 crc kubenswrapper[4824]: I1209 10:15:56.909267 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.112730 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-ceilometer-tls-certs\") pod \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.112769 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dlhv\" (UniqueName: \"kubernetes.io/projected/d254951a-1dd1-4b09-8cfa-e765853cb5e9-kube-api-access-9dlhv\") pod \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.112821 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-sg-core-conf-yaml\") pod \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.113103 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-combined-ca-bundle\") pod \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.113284 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d254951a-1dd1-4b09-8cfa-e765853cb5e9-run-httpd\") pod \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.113352 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d254951a-1dd1-4b09-8cfa-e765853cb5e9-log-httpd\") pod \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.113500 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-scripts\") pod \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.113659 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-config-data\") pod \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\" (UID: \"d254951a-1dd1-4b09-8cfa-e765853cb5e9\") " Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.114123 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d254951a-1dd1-4b09-8cfa-e765853cb5e9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d254951a-1dd1-4b09-8cfa-e765853cb5e9" (UID: "d254951a-1dd1-4b09-8cfa-e765853cb5e9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.114262 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d254951a-1dd1-4b09-8cfa-e765853cb5e9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d254951a-1dd1-4b09-8cfa-e765853cb5e9" (UID: "d254951a-1dd1-4b09-8cfa-e765853cb5e9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.114951 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d254951a-1dd1-4b09-8cfa-e765853cb5e9-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.114979 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d254951a-1dd1-4b09-8cfa-e765853cb5e9-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.119875 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d254951a-1dd1-4b09-8cfa-e765853cb5e9-kube-api-access-9dlhv" (OuterVolumeSpecName: "kube-api-access-9dlhv") pod "d254951a-1dd1-4b09-8cfa-e765853cb5e9" (UID: "d254951a-1dd1-4b09-8cfa-e765853cb5e9"). InnerVolumeSpecName "kube-api-access-9dlhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.120667 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-scripts" (OuterVolumeSpecName: "scripts") pod "d254951a-1dd1-4b09-8cfa-e765853cb5e9" (UID: "d254951a-1dd1-4b09-8cfa-e765853cb5e9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.152508 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d254951a-1dd1-4b09-8cfa-e765853cb5e9" (UID: "d254951a-1dd1-4b09-8cfa-e765853cb5e9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.177861 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "d254951a-1dd1-4b09-8cfa-e765853cb5e9" (UID: "d254951a-1dd1-4b09-8cfa-e765853cb5e9"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.216161 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.216197 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.216208 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dlhv\" (UniqueName: \"kubernetes.io/projected/d254951a-1dd1-4b09-8cfa-e765853cb5e9-kube-api-access-9dlhv\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.216220 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.223944 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d254951a-1dd1-4b09-8cfa-e765853cb5e9" (UID: "d254951a-1dd1-4b09-8cfa-e765853cb5e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.244159 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-config-data" (OuterVolumeSpecName: "config-data") pod "d254951a-1dd1-4b09-8cfa-e765853cb5e9" (UID: "d254951a-1dd1-4b09-8cfa-e765853cb5e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.318649 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.318685 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d254951a-1dd1-4b09-8cfa-e765853cb5e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.443172 4824 generic.go:334] "Generic (PLEG): container finished" podID="d254951a-1dd1-4b09-8cfa-e765853cb5e9" containerID="a2f5f62a717870a67acb6dd5fc736f75d1a2b516521177079b70d23cdbe5ce0f" exitCode=137 Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.443255 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d254951a-1dd1-4b09-8cfa-e765853cb5e9","Type":"ContainerDied","Data":"a2f5f62a717870a67acb6dd5fc736f75d1a2b516521177079b70d23cdbe5ce0f"} Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.443308 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d254951a-1dd1-4b09-8cfa-e765853cb5e9","Type":"ContainerDied","Data":"4f8718a387b02f1a62c3272b0b6fc52b6dd12ac62f11c441842290a0d4102a51"} Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.443970 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.443420 4824 scope.go:117] "RemoveContainer" containerID="a2f5f62a717870a67acb6dd5fc736f75d1a2b516521177079b70d23cdbe5ce0f" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.443310 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.457706 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.538408 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.543947 4824 scope.go:117] "RemoveContainer" containerID="6ecc53fd2da09ec596a0864786ec17e2f907598c43db35eabbe9f964824bb708" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.562356 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.696672 4824 scope.go:117] "RemoveContainer" containerID="aa021504dd953ba53062645025462a4f01abb00b32816d01f0196aaa2ade5b7d" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.704553 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:15:57 crc kubenswrapper[4824]: E1209 10:15:57.705107 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d254951a-1dd1-4b09-8cfa-e765853cb5e9" containerName="proxy-httpd" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.705129 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d254951a-1dd1-4b09-8cfa-e765853cb5e9" containerName="proxy-httpd" Dec 09 10:15:57 crc kubenswrapper[4824]: E1209 10:15:57.705160 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d254951a-1dd1-4b09-8cfa-e765853cb5e9" containerName="sg-core" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.705167 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d254951a-1dd1-4b09-8cfa-e765853cb5e9" containerName="sg-core" Dec 09 10:15:57 crc kubenswrapper[4824]: E1209 10:15:57.705180 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d254951a-1dd1-4b09-8cfa-e765853cb5e9" containerName="ceilometer-notification-agent" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.705208 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d254951a-1dd1-4b09-8cfa-e765853cb5e9" containerName="ceilometer-notification-agent" Dec 09 10:15:57 crc kubenswrapper[4824]: E1209 10:15:57.705226 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d254951a-1dd1-4b09-8cfa-e765853cb5e9" containerName="ceilometer-central-agent" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.705232 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d254951a-1dd1-4b09-8cfa-e765853cb5e9" containerName="ceilometer-central-agent" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.705496 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d254951a-1dd1-4b09-8cfa-e765853cb5e9" containerName="ceilometer-central-agent" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.705517 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d254951a-1dd1-4b09-8cfa-e765853cb5e9" containerName="proxy-httpd" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.705548 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d254951a-1dd1-4b09-8cfa-e765853cb5e9" containerName="sg-core" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.705573 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d254951a-1dd1-4b09-8cfa-e765853cb5e9" containerName="ceilometer-notification-agent" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.709168 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.720795 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.721682 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.722534 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.742431 4824 scope.go:117] "RemoveContainer" containerID="1cd4211ca88d7f9721e9b7c47366ab131e71cdd2d1bc02648cdec6c0c0719677" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.746353 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.771436 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.771524 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-scripts\") pod \"ceilometer-0\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.771653 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.771929 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-run-httpd\") pod \"ceilometer-0\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.772040 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-log-httpd\") pod \"ceilometer-0\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.772059 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-config-data\") pod \"ceilometer-0\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.772136 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6fbg\" (UniqueName: \"kubernetes.io/projected/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-kube-api-access-w6fbg\") pod \"ceilometer-0\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.772299 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.790061 4824 scope.go:117] "RemoveContainer" containerID="a2f5f62a717870a67acb6dd5fc736f75d1a2b516521177079b70d23cdbe5ce0f" Dec 09 10:15:57 crc kubenswrapper[4824]: E1209 10:15:57.790652 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2f5f62a717870a67acb6dd5fc736f75d1a2b516521177079b70d23cdbe5ce0f\": container with ID starting with a2f5f62a717870a67acb6dd5fc736f75d1a2b516521177079b70d23cdbe5ce0f not found: ID does not exist" containerID="a2f5f62a717870a67acb6dd5fc736f75d1a2b516521177079b70d23cdbe5ce0f" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.790711 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2f5f62a717870a67acb6dd5fc736f75d1a2b516521177079b70d23cdbe5ce0f"} err="failed to get container status \"a2f5f62a717870a67acb6dd5fc736f75d1a2b516521177079b70d23cdbe5ce0f\": rpc error: code = NotFound desc = could not find container \"a2f5f62a717870a67acb6dd5fc736f75d1a2b516521177079b70d23cdbe5ce0f\": container with ID starting with a2f5f62a717870a67acb6dd5fc736f75d1a2b516521177079b70d23cdbe5ce0f not found: ID does not exist" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.790741 4824 scope.go:117] "RemoveContainer" containerID="6ecc53fd2da09ec596a0864786ec17e2f907598c43db35eabbe9f964824bb708" Dec 09 10:15:57 crc kubenswrapper[4824]: E1209 10:15:57.791302 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ecc53fd2da09ec596a0864786ec17e2f907598c43db35eabbe9f964824bb708\": container with ID starting with 6ecc53fd2da09ec596a0864786ec17e2f907598c43db35eabbe9f964824bb708 not found: ID does not exist" containerID="6ecc53fd2da09ec596a0864786ec17e2f907598c43db35eabbe9f964824bb708" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.791352 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ecc53fd2da09ec596a0864786ec17e2f907598c43db35eabbe9f964824bb708"} err="failed to get container status \"6ecc53fd2da09ec596a0864786ec17e2f907598c43db35eabbe9f964824bb708\": rpc error: code = NotFound desc = could not find container \"6ecc53fd2da09ec596a0864786ec17e2f907598c43db35eabbe9f964824bb708\": container with ID starting with 6ecc53fd2da09ec596a0864786ec17e2f907598c43db35eabbe9f964824bb708 not found: ID does not exist" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.791379 4824 scope.go:117] "RemoveContainer" containerID="aa021504dd953ba53062645025462a4f01abb00b32816d01f0196aaa2ade5b7d" Dec 09 10:15:57 crc kubenswrapper[4824]: E1209 10:15:57.791673 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa021504dd953ba53062645025462a4f01abb00b32816d01f0196aaa2ade5b7d\": container with ID starting with aa021504dd953ba53062645025462a4f01abb00b32816d01f0196aaa2ade5b7d not found: ID does not exist" containerID="aa021504dd953ba53062645025462a4f01abb00b32816d01f0196aaa2ade5b7d" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.791703 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa021504dd953ba53062645025462a4f01abb00b32816d01f0196aaa2ade5b7d"} err="failed to get container status \"aa021504dd953ba53062645025462a4f01abb00b32816d01f0196aaa2ade5b7d\": rpc error: code = NotFound desc = could not find container \"aa021504dd953ba53062645025462a4f01abb00b32816d01f0196aaa2ade5b7d\": container with ID starting with aa021504dd953ba53062645025462a4f01abb00b32816d01f0196aaa2ade5b7d not found: ID does not exist" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.791719 4824 scope.go:117] "RemoveContainer" containerID="1cd4211ca88d7f9721e9b7c47366ab131e71cdd2d1bc02648cdec6c0c0719677" Dec 09 10:15:57 crc kubenswrapper[4824]: E1209 10:15:57.792041 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cd4211ca88d7f9721e9b7c47366ab131e71cdd2d1bc02648cdec6c0c0719677\": container with ID starting with 1cd4211ca88d7f9721e9b7c47366ab131e71cdd2d1bc02648cdec6c0c0719677 not found: ID does not exist" containerID="1cd4211ca88d7f9721e9b7c47366ab131e71cdd2d1bc02648cdec6c0c0719677" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.792076 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cd4211ca88d7f9721e9b7c47366ab131e71cdd2d1bc02648cdec6c0c0719677"} err="failed to get container status \"1cd4211ca88d7f9721e9b7c47366ab131e71cdd2d1bc02648cdec6c0c0719677\": rpc error: code = NotFound desc = could not find container \"1cd4211ca88d7f9721e9b7c47366ab131e71cdd2d1bc02648cdec6c0c0719677\": container with ID starting with 1cd4211ca88d7f9721e9b7c47366ab131e71cdd2d1bc02648cdec6c0c0719677 not found: ID does not exist" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.874429 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-log-httpd\") pod \"ceilometer-0\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.874485 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-config-data\") pod \"ceilometer-0\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.874520 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6fbg\" (UniqueName: \"kubernetes.io/projected/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-kube-api-access-w6fbg\") pod \"ceilometer-0\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.874617 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.874675 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.874721 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-scripts\") pod \"ceilometer-0\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.874745 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.874825 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-run-httpd\") pod \"ceilometer-0\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.874903 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-log-httpd\") pod \"ceilometer-0\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.875137 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-run-httpd\") pod \"ceilometer-0\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.878978 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.879064 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-scripts\") pod \"ceilometer-0\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.880505 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.886599 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.887222 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-config-data\") pod \"ceilometer-0\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.900540 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6fbg\" (UniqueName: \"kubernetes.io/projected/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-kube-api-access-w6fbg\") pod \"ceilometer-0\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " pod="openstack/ceilometer-0" Dec 09 10:15:57 crc kubenswrapper[4824]: I1209 10:15:57.925810 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d254951a-1dd1-4b09-8cfa-e765853cb5e9" path="/var/lib/kubelet/pods/d254951a-1dd1-4b09-8cfa-e765853cb5e9/volumes" Dec 09 10:15:58 crc kubenswrapper[4824]: I1209 10:15:58.035696 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:15:58 crc kubenswrapper[4824]: I1209 10:15:58.525588 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:15:59 crc kubenswrapper[4824]: I1209 10:15:59.482637 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31b5f1b5-0f60-47f4-b4a6-6b457ea622be","Type":"ContainerStarted","Data":"066750eead3e694d0dc6b7ddabcbde71a628f40e58f98c45727562ff2269f30c"} Dec 09 10:15:59 crc kubenswrapper[4824]: I1209 10:15:59.483137 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31b5f1b5-0f60-47f4-b4a6-6b457ea622be","Type":"ContainerStarted","Data":"17bbefb3bd826345b125982f0265e211abd904ee3e6caefed64c4562ddd90c9a"} Dec 09 10:15:59 crc kubenswrapper[4824]: I1209 10:15:59.911058 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:15:59 crc kubenswrapper[4824]: E1209 10:15:59.911503 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:16:00 crc kubenswrapper[4824]: I1209 10:16:00.502047 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31b5f1b5-0f60-47f4-b4a6-6b457ea622be","Type":"ContainerStarted","Data":"31332a7c9ba75e6a67f37596510b470eb420a7ffcfbea609a3ac1efd24e7a3ec"} Dec 09 10:16:00 crc kubenswrapper[4824]: I1209 10:16:00.851351 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 09 10:16:00 crc kubenswrapper[4824]: I1209 10:16:00.853557 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 09 10:16:00 crc kubenswrapper[4824]: I1209 10:16:00.861433 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 09 10:16:01 crc kubenswrapper[4824]: I1209 10:16:01.518896 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31b5f1b5-0f60-47f4-b4a6-6b457ea622be","Type":"ContainerStarted","Data":"b42b3ca64da6ae1b9dba2cc35d19d60ff830c508b053eef56824f66ff004c329"} Dec 09 10:16:01 crc kubenswrapper[4824]: I1209 10:16:01.527088 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 09 10:16:03 crc kubenswrapper[4824]: I1209 10:16:03.575524 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31b5f1b5-0f60-47f4-b4a6-6b457ea622be","Type":"ContainerStarted","Data":"c9d596aa7c77e733aff0fa6a7db404be9deaa883f07febf1dac4160a367c0c6e"} Dec 09 10:16:03 crc kubenswrapper[4824]: I1209 10:16:03.600362 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.544889471 podStartE2EDuration="6.600338669s" podCreationTimestamp="2025-12-09 10:15:57 +0000 UTC" firstStartedPulling="2025-12-09 10:15:58.528542137 +0000 UTC m=+1714.863046804" lastFinishedPulling="2025-12-09 10:16:02.583991335 +0000 UTC m=+1718.918496002" observedRunningTime="2025-12-09 10:16:03.59974035 +0000 UTC m=+1719.934245017" watchObservedRunningTime="2025-12-09 10:16:03.600338669 +0000 UTC m=+1719.934843356" Dec 09 10:16:04 crc kubenswrapper[4824]: I1209 10:16:04.586941 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 10:16:12 crc kubenswrapper[4824]: I1209 10:16:12.942495 4824 scope.go:117] "RemoveContainer" containerID="0a579075bdc0f1dc087391998446e9f4e87f86205adca94cd10a32f6dcd5b13b" Dec 09 10:16:12 crc kubenswrapper[4824]: I1209 10:16:12.988806 4824 scope.go:117] "RemoveContainer" containerID="99db94c0c3cfc2e1ecab46d0a7b0c3c3d0fcc1067806b3f5b0b09fba74941434" Dec 09 10:16:13 crc kubenswrapper[4824]: I1209 10:16:13.060151 4824 scope.go:117] "RemoveContainer" containerID="065fe900e020f72eec27757202af458fb43910dff8bc9cca1e8b163284ff096f" Dec 09 10:16:13 crc kubenswrapper[4824]: I1209 10:16:13.083002 4824 scope.go:117] "RemoveContainer" containerID="3bcac543b2d89d00cda20308eec14bbc4383f0f9d93d98669cd2beae0742c17f" Dec 09 10:16:13 crc kubenswrapper[4824]: I1209 10:16:13.921447 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:16:13 crc kubenswrapper[4824]: E1209 10:16:13.922353 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:16:24 crc kubenswrapper[4824]: I1209 10:16:24.910874 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:16:24 crc kubenswrapper[4824]: E1209 10:16:24.911710 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:16:28 crc kubenswrapper[4824]: I1209 10:16:28.049376 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 09 10:16:37 crc kubenswrapper[4824]: I1209 10:16:37.910399 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:16:37 crc kubenswrapper[4824]: E1209 10:16:37.911248 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:16:41 crc kubenswrapper[4824]: I1209 10:16:41.666008 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-vfnsw"] Dec 09 10:16:41 crc kubenswrapper[4824]: I1209 10:16:41.677010 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-vfnsw"] Dec 09 10:16:41 crc kubenswrapper[4824]: I1209 10:16:41.761273 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-m6t59"] Dec 09 10:16:41 crc kubenswrapper[4824]: I1209 10:16:41.763359 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-m6t59" Dec 09 10:16:41 crc kubenswrapper[4824]: I1209 10:16:41.773727 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-m6t59"] Dec 09 10:16:41 crc kubenswrapper[4824]: I1209 10:16:41.875564 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8051e79f-a20f-4a10-b2f8-d8c770328797-config-data\") pod \"heat-db-sync-m6t59\" (UID: \"8051e79f-a20f-4a10-b2f8-d8c770328797\") " pod="openstack/heat-db-sync-m6t59" Dec 09 10:16:41 crc kubenswrapper[4824]: I1209 10:16:41.875683 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8051e79f-a20f-4a10-b2f8-d8c770328797-combined-ca-bundle\") pod \"heat-db-sync-m6t59\" (UID: \"8051e79f-a20f-4a10-b2f8-d8c770328797\") " pod="openstack/heat-db-sync-m6t59" Dec 09 10:16:41 crc kubenswrapper[4824]: I1209 10:16:41.875765 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2pzz\" (UniqueName: \"kubernetes.io/projected/8051e79f-a20f-4a10-b2f8-d8c770328797-kube-api-access-r2pzz\") pod \"heat-db-sync-m6t59\" (UID: \"8051e79f-a20f-4a10-b2f8-d8c770328797\") " pod="openstack/heat-db-sync-m6t59" Dec 09 10:16:41 crc kubenswrapper[4824]: I1209 10:16:41.925773 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="405d0a5e-fd62-4146-bfb9-96ff6c077836" path="/var/lib/kubelet/pods/405d0a5e-fd62-4146-bfb9-96ff6c077836/volumes" Dec 09 10:16:41 crc kubenswrapper[4824]: I1209 10:16:41.977989 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8051e79f-a20f-4a10-b2f8-d8c770328797-combined-ca-bundle\") pod \"heat-db-sync-m6t59\" (UID: \"8051e79f-a20f-4a10-b2f8-d8c770328797\") " pod="openstack/heat-db-sync-m6t59" Dec 09 10:16:41 crc kubenswrapper[4824]: I1209 10:16:41.978164 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2pzz\" (UniqueName: \"kubernetes.io/projected/8051e79f-a20f-4a10-b2f8-d8c770328797-kube-api-access-r2pzz\") pod \"heat-db-sync-m6t59\" (UID: \"8051e79f-a20f-4a10-b2f8-d8c770328797\") " pod="openstack/heat-db-sync-m6t59" Dec 09 10:16:41 crc kubenswrapper[4824]: I1209 10:16:41.978356 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8051e79f-a20f-4a10-b2f8-d8c770328797-config-data\") pod \"heat-db-sync-m6t59\" (UID: \"8051e79f-a20f-4a10-b2f8-d8c770328797\") " pod="openstack/heat-db-sync-m6t59" Dec 09 10:16:41 crc kubenswrapper[4824]: I1209 10:16:41.984797 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8051e79f-a20f-4a10-b2f8-d8c770328797-combined-ca-bundle\") pod \"heat-db-sync-m6t59\" (UID: \"8051e79f-a20f-4a10-b2f8-d8c770328797\") " pod="openstack/heat-db-sync-m6t59" Dec 09 10:16:41 crc kubenswrapper[4824]: I1209 10:16:41.986849 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8051e79f-a20f-4a10-b2f8-d8c770328797-config-data\") pod \"heat-db-sync-m6t59\" (UID: \"8051e79f-a20f-4a10-b2f8-d8c770328797\") " pod="openstack/heat-db-sync-m6t59" Dec 09 10:16:41 crc kubenswrapper[4824]: I1209 10:16:41.999712 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2pzz\" (UniqueName: \"kubernetes.io/projected/8051e79f-a20f-4a10-b2f8-d8c770328797-kube-api-access-r2pzz\") pod \"heat-db-sync-m6t59\" (UID: \"8051e79f-a20f-4a10-b2f8-d8c770328797\") " pod="openstack/heat-db-sync-m6t59" Dec 09 10:16:42 crc kubenswrapper[4824]: I1209 10:16:42.102748 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-m6t59" Dec 09 10:16:42 crc kubenswrapper[4824]: I1209 10:16:42.629156 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-m6t59"] Dec 09 10:16:43 crc kubenswrapper[4824]: I1209 10:16:43.120847 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-m6t59" event={"ID":"8051e79f-a20f-4a10-b2f8-d8c770328797","Type":"ContainerStarted","Data":"3b3174fe8553a469c25ed20745542d8c596a60c10b553d1cabcba892ef1c9060"} Dec 09 10:16:43 crc kubenswrapper[4824]: I1209 10:16:43.783113 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:16:43 crc kubenswrapper[4824]: I1209 10:16:43.783649 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="31b5f1b5-0f60-47f4-b4a6-6b457ea622be" containerName="ceilometer-central-agent" containerID="cri-o://066750eead3e694d0dc6b7ddabcbde71a628f40e58f98c45727562ff2269f30c" gracePeriod=30 Dec 09 10:16:43 crc kubenswrapper[4824]: I1209 10:16:43.784387 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="31b5f1b5-0f60-47f4-b4a6-6b457ea622be" containerName="proxy-httpd" containerID="cri-o://c9d596aa7c77e733aff0fa6a7db404be9deaa883f07febf1dac4160a367c0c6e" gracePeriod=30 Dec 09 10:16:43 crc kubenswrapper[4824]: I1209 10:16:43.784445 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="31b5f1b5-0f60-47f4-b4a6-6b457ea622be" containerName="sg-core" containerID="cri-o://b42b3ca64da6ae1b9dba2cc35d19d60ff830c508b053eef56824f66ff004c329" gracePeriod=30 Dec 09 10:16:43 crc kubenswrapper[4824]: I1209 10:16:43.784478 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="31b5f1b5-0f60-47f4-b4a6-6b457ea622be" containerName="ceilometer-notification-agent" containerID="cri-o://31332a7c9ba75e6a67f37596510b470eb420a7ffcfbea609a3ac1efd24e7a3ec" gracePeriod=30 Dec 09 10:16:44 crc kubenswrapper[4824]: I1209 10:16:44.139893 4824 generic.go:334] "Generic (PLEG): container finished" podID="31b5f1b5-0f60-47f4-b4a6-6b457ea622be" containerID="c9d596aa7c77e733aff0fa6a7db404be9deaa883f07febf1dac4160a367c0c6e" exitCode=0 Dec 09 10:16:44 crc kubenswrapper[4824]: I1209 10:16:44.139933 4824 generic.go:334] "Generic (PLEG): container finished" podID="31b5f1b5-0f60-47f4-b4a6-6b457ea622be" containerID="b42b3ca64da6ae1b9dba2cc35d19d60ff830c508b053eef56824f66ff004c329" exitCode=2 Dec 09 10:16:44 crc kubenswrapper[4824]: I1209 10:16:44.139924 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31b5f1b5-0f60-47f4-b4a6-6b457ea622be","Type":"ContainerDied","Data":"c9d596aa7c77e733aff0fa6a7db404be9deaa883f07febf1dac4160a367c0c6e"} Dec 09 10:16:44 crc kubenswrapper[4824]: I1209 10:16:44.139973 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31b5f1b5-0f60-47f4-b4a6-6b457ea622be","Type":"ContainerDied","Data":"b42b3ca64da6ae1b9dba2cc35d19d60ff830c508b053eef56824f66ff004c329"} Dec 09 10:16:44 crc kubenswrapper[4824]: I1209 10:16:44.692418 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-2"] Dec 09 10:16:45 crc kubenswrapper[4824]: I1209 10:16:45.166378 4824 generic.go:334] "Generic (PLEG): container finished" podID="31b5f1b5-0f60-47f4-b4a6-6b457ea622be" containerID="066750eead3e694d0dc6b7ddabcbde71a628f40e58f98c45727562ff2269f30c" exitCode=0 Dec 09 10:16:45 crc kubenswrapper[4824]: I1209 10:16:45.166413 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31b5f1b5-0f60-47f4-b4a6-6b457ea622be","Type":"ContainerDied","Data":"066750eead3e694d0dc6b7ddabcbde71a628f40e58f98c45727562ff2269f30c"} Dec 09 10:16:46 crc kubenswrapper[4824]: I1209 10:16:46.010117 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 10:16:49 crc kubenswrapper[4824]: I1209 10:16:49.227820 4824 generic.go:334] "Generic (PLEG): container finished" podID="31b5f1b5-0f60-47f4-b4a6-6b457ea622be" containerID="31332a7c9ba75e6a67f37596510b470eb420a7ffcfbea609a3ac1efd24e7a3ec" exitCode=0 Dec 09 10:16:49 crc kubenswrapper[4824]: I1209 10:16:49.228063 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31b5f1b5-0f60-47f4-b4a6-6b457ea622be","Type":"ContainerDied","Data":"31332a7c9ba75e6a67f37596510b470eb420a7ffcfbea609a3ac1efd24e7a3ec"} Dec 09 10:16:49 crc kubenswrapper[4824]: I1209 10:16:49.236094 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-2" podUID="09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" containerName="rabbitmq" containerID="cri-o://3ad239061116629b299b08f30e913aae04ed870beb2645bdf1ab9cfc59480f62" gracePeriod=604796 Dec 09 10:16:50 crc kubenswrapper[4824]: I1209 10:16:50.806093 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="07c1822c-1257-4d8c-9a17-d921929000fa" containerName="rabbitmq" containerID="cri-o://d35277030f4a65e4facfc048ac021fc3c08167f2f2aceae17375fa39c64384b6" gracePeriod=604796 Dec 09 10:16:52 crc kubenswrapper[4824]: I1209 10:16:52.912479 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:16:52 crc kubenswrapper[4824]: E1209 10:16:52.913528 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:16:55 crc kubenswrapper[4824]: I1209 10:16:55.929869 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-2" podUID="09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.131:5671: connect: connection refused" Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.336900 4824 generic.go:334] "Generic (PLEG): container finished" podID="09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" containerID="3ad239061116629b299b08f30e913aae04ed870beb2645bdf1ab9cfc59480f62" exitCode=0 Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.337017 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55","Type":"ContainerDied","Data":"3ad239061116629b299b08f30e913aae04ed870beb2645bdf1ab9cfc59480f62"} Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.343262 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31b5f1b5-0f60-47f4-b4a6-6b457ea622be","Type":"ContainerDied","Data":"17bbefb3bd826345b125982f0265e211abd904ee3e6caefed64c4562ddd90c9a"} Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.343306 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17bbefb3bd826345b125982f0265e211abd904ee3e6caefed64c4562ddd90c9a" Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.454668 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.610264 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-ceilometer-tls-certs\") pod \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.610452 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-combined-ca-bundle\") pod \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.610489 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-sg-core-conf-yaml\") pod \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.610573 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-log-httpd\") pod \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.610664 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-run-httpd\") pod \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.610741 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-config-data\") pod \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.610818 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-scripts\") pod \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.610852 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6fbg\" (UniqueName: \"kubernetes.io/projected/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-kube-api-access-w6fbg\") pod \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\" (UID: \"31b5f1b5-0f60-47f4-b4a6-6b457ea622be\") " Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.620544 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "31b5f1b5-0f60-47f4-b4a6-6b457ea622be" (UID: "31b5f1b5-0f60-47f4-b4a6-6b457ea622be"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.620827 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "31b5f1b5-0f60-47f4-b4a6-6b457ea622be" (UID: "31b5f1b5-0f60-47f4-b4a6-6b457ea622be"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.636570 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-scripts" (OuterVolumeSpecName: "scripts") pod "31b5f1b5-0f60-47f4-b4a6-6b457ea622be" (UID: "31b5f1b5-0f60-47f4-b4a6-6b457ea622be"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.636665 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-kube-api-access-w6fbg" (OuterVolumeSpecName: "kube-api-access-w6fbg") pod "31b5f1b5-0f60-47f4-b4a6-6b457ea622be" (UID: "31b5f1b5-0f60-47f4-b4a6-6b457ea622be"). InnerVolumeSpecName "kube-api-access-w6fbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.684919 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "31b5f1b5-0f60-47f4-b4a6-6b457ea622be" (UID: "31b5f1b5-0f60-47f4-b4a6-6b457ea622be"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.687203 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "31b5f1b5-0f60-47f4-b4a6-6b457ea622be" (UID: "31b5f1b5-0f60-47f4-b4a6-6b457ea622be"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.726117 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.726162 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.726174 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.726185 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.726200 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6fbg\" (UniqueName: \"kubernetes.io/projected/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-kube-api-access-w6fbg\") on node \"crc\" DevicePath \"\"" Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.726213 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.790894 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "31b5f1b5-0f60-47f4-b4a6-6b457ea622be" (UID: "31b5f1b5-0f60-47f4-b4a6-6b457ea622be"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.816730 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-config-data" (OuterVolumeSpecName: "config-data") pod "31b5f1b5-0f60-47f4-b4a6-6b457ea622be" (UID: "31b5f1b5-0f60-47f4-b4a6-6b457ea622be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.832818 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:16:56 crc kubenswrapper[4824]: I1209 10:16:56.833232 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31b5f1b5-0f60-47f4-b4a6-6b457ea622be-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.130566 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="07c1822c-1257-4d8c-9a17-d921929000fa" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.132:5671: connect: connection refused" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.372478 4824 generic.go:334] "Generic (PLEG): container finished" podID="07c1822c-1257-4d8c-9a17-d921929000fa" containerID="d35277030f4a65e4facfc048ac021fc3c08167f2f2aceae17375fa39c64384b6" exitCode=0 Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.372937 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.372808 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"07c1822c-1257-4d8c-9a17-d921929000fa","Type":"ContainerDied","Data":"d35277030f4a65e4facfc048ac021fc3c08167f2f2aceae17375fa39c64384b6"} Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.462091 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.474613 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.500198 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:16:57 crc kubenswrapper[4824]: E1209 10:16:57.500840 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31b5f1b5-0f60-47f4-b4a6-6b457ea622be" containerName="sg-core" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.500863 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="31b5f1b5-0f60-47f4-b4a6-6b457ea622be" containerName="sg-core" Dec 09 10:16:57 crc kubenswrapper[4824]: E1209 10:16:57.500897 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31b5f1b5-0f60-47f4-b4a6-6b457ea622be" containerName="proxy-httpd" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.500906 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="31b5f1b5-0f60-47f4-b4a6-6b457ea622be" containerName="proxy-httpd" Dec 09 10:16:57 crc kubenswrapper[4824]: E1209 10:16:57.500930 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31b5f1b5-0f60-47f4-b4a6-6b457ea622be" containerName="ceilometer-notification-agent" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.500940 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="31b5f1b5-0f60-47f4-b4a6-6b457ea622be" containerName="ceilometer-notification-agent" Dec 09 10:16:57 crc kubenswrapper[4824]: E1209 10:16:57.500961 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31b5f1b5-0f60-47f4-b4a6-6b457ea622be" containerName="ceilometer-central-agent" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.500969 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="31b5f1b5-0f60-47f4-b4a6-6b457ea622be" containerName="ceilometer-central-agent" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.501293 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="31b5f1b5-0f60-47f4-b4a6-6b457ea622be" containerName="proxy-httpd" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.501325 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="31b5f1b5-0f60-47f4-b4a6-6b457ea622be" containerName="ceilometer-central-agent" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.501358 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="31b5f1b5-0f60-47f4-b4a6-6b457ea622be" containerName="ceilometer-notification-agent" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.501371 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="31b5f1b5-0f60-47f4-b4a6-6b457ea622be" containerName="sg-core" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.508128 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.511430 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.511717 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.511922 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.537718 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.557244 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p7fx\" (UniqueName: \"kubernetes.io/projected/9339c10c-1cea-4ffc-8931-9b349a1e74cd-kube-api-access-6p7fx\") pod \"ceilometer-0\" (UID: \"9339c10c-1cea-4ffc-8931-9b349a1e74cd\") " pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.557354 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9339c10c-1cea-4ffc-8931-9b349a1e74cd-config-data\") pod \"ceilometer-0\" (UID: \"9339c10c-1cea-4ffc-8931-9b349a1e74cd\") " pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.557373 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9339c10c-1cea-4ffc-8931-9b349a1e74cd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9339c10c-1cea-4ffc-8931-9b349a1e74cd\") " pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.557453 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9339c10c-1cea-4ffc-8931-9b349a1e74cd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9339c10c-1cea-4ffc-8931-9b349a1e74cd\") " pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.557475 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9339c10c-1cea-4ffc-8931-9b349a1e74cd-scripts\") pod \"ceilometer-0\" (UID: \"9339c10c-1cea-4ffc-8931-9b349a1e74cd\") " pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.557490 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9339c10c-1cea-4ffc-8931-9b349a1e74cd-run-httpd\") pod \"ceilometer-0\" (UID: \"9339c10c-1cea-4ffc-8931-9b349a1e74cd\") " pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.557592 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9339c10c-1cea-4ffc-8931-9b349a1e74cd-log-httpd\") pod \"ceilometer-0\" (UID: \"9339c10c-1cea-4ffc-8931-9b349a1e74cd\") " pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.557828 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9339c10c-1cea-4ffc-8931-9b349a1e74cd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9339c10c-1cea-4ffc-8931-9b349a1e74cd\") " pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.661847 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9339c10c-1cea-4ffc-8931-9b349a1e74cd-config-data\") pod \"ceilometer-0\" (UID: \"9339c10c-1cea-4ffc-8931-9b349a1e74cd\") " pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.661926 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9339c10c-1cea-4ffc-8931-9b349a1e74cd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9339c10c-1cea-4ffc-8931-9b349a1e74cd\") " pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.662010 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9339c10c-1cea-4ffc-8931-9b349a1e74cd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9339c10c-1cea-4ffc-8931-9b349a1e74cd\") " pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.662063 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9339c10c-1cea-4ffc-8931-9b349a1e74cd-scripts\") pod \"ceilometer-0\" (UID: \"9339c10c-1cea-4ffc-8931-9b349a1e74cd\") " pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.662085 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9339c10c-1cea-4ffc-8931-9b349a1e74cd-run-httpd\") pod \"ceilometer-0\" (UID: \"9339c10c-1cea-4ffc-8931-9b349a1e74cd\") " pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.662226 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9339c10c-1cea-4ffc-8931-9b349a1e74cd-log-httpd\") pod \"ceilometer-0\" (UID: \"9339c10c-1cea-4ffc-8931-9b349a1e74cd\") " pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.662422 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9339c10c-1cea-4ffc-8931-9b349a1e74cd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9339c10c-1cea-4ffc-8931-9b349a1e74cd\") " pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.662988 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p7fx\" (UniqueName: \"kubernetes.io/projected/9339c10c-1cea-4ffc-8931-9b349a1e74cd-kube-api-access-6p7fx\") pod \"ceilometer-0\" (UID: \"9339c10c-1cea-4ffc-8931-9b349a1e74cd\") " pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.663277 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9339c10c-1cea-4ffc-8931-9b349a1e74cd-log-httpd\") pod \"ceilometer-0\" (UID: \"9339c10c-1cea-4ffc-8931-9b349a1e74cd\") " pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.663640 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9339c10c-1cea-4ffc-8931-9b349a1e74cd-run-httpd\") pod \"ceilometer-0\" (UID: \"9339c10c-1cea-4ffc-8931-9b349a1e74cd\") " pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.668965 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9339c10c-1cea-4ffc-8931-9b349a1e74cd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9339c10c-1cea-4ffc-8931-9b349a1e74cd\") " pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.668990 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9339c10c-1cea-4ffc-8931-9b349a1e74cd-scripts\") pod \"ceilometer-0\" (UID: \"9339c10c-1cea-4ffc-8931-9b349a1e74cd\") " pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.669313 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9339c10c-1cea-4ffc-8931-9b349a1e74cd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9339c10c-1cea-4ffc-8931-9b349a1e74cd\") " pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.669544 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9339c10c-1cea-4ffc-8931-9b349a1e74cd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9339c10c-1cea-4ffc-8931-9b349a1e74cd\") " pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.672291 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9339c10c-1cea-4ffc-8931-9b349a1e74cd-config-data\") pod \"ceilometer-0\" (UID: \"9339c10c-1cea-4ffc-8931-9b349a1e74cd\") " pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.682157 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p7fx\" (UniqueName: \"kubernetes.io/projected/9339c10c-1cea-4ffc-8931-9b349a1e74cd-kube-api-access-6p7fx\") pod \"ceilometer-0\" (UID: \"9339c10c-1cea-4ffc-8931-9b349a1e74cd\") " pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.839250 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 10:16:57 crc kubenswrapper[4824]: I1209 10:16:57.934224 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31b5f1b5-0f60-47f4-b4a6-6b457ea622be" path="/var/lib/kubelet/pods/31b5f1b5-0f60-47f4-b4a6-6b457ea622be/volumes" Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.378426 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-x2dv9"] Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.382135 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.385063 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.402954 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-x2dv9"] Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.486693 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-x2dv9\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.487026 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-x2dv9\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.487056 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-x2dv9\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.487101 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-x2dv9\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.487426 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-x2dv9\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.487475 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqmf9\" (UniqueName: \"kubernetes.io/projected/9ce40f74-fec2-4b15-bb84-d8972d806882-kube-api-access-tqmf9\") pod \"dnsmasq-dns-7d84b4d45c-x2dv9\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.487554 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-config\") pod \"dnsmasq-dns-7d84b4d45c-x2dv9\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.590402 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-x2dv9\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.590488 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-x2dv9\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.590522 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-x2dv9\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.590584 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-x2dv9\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.590714 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqmf9\" (UniqueName: \"kubernetes.io/projected/9ce40f74-fec2-4b15-bb84-d8972d806882-kube-api-access-tqmf9\") pod \"dnsmasq-dns-7d84b4d45c-x2dv9\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.590740 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-x2dv9\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.590799 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-config\") pod \"dnsmasq-dns-7d84b4d45c-x2dv9\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.591621 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-x2dv9\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.591677 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-config\") pod \"dnsmasq-dns-7d84b4d45c-x2dv9\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.591921 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-x2dv9\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.592400 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-x2dv9\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.592596 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-x2dv9\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.592891 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-x2dv9\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.628703 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqmf9\" (UniqueName: \"kubernetes.io/projected/9ce40f74-fec2-4b15-bb84-d8972d806882-kube-api-access-tqmf9\") pod \"dnsmasq-dns-7d84b4d45c-x2dv9\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:01 crc kubenswrapper[4824]: I1209 10:17:01.725916 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:05 crc kubenswrapper[4824]: I1209 10:17:05.915354 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:17:05 crc kubenswrapper[4824]: E1209 10:17:05.916114 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:17:06 crc kubenswrapper[4824]: E1209 10:17:06.726801 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 10:17:06 crc kubenswrapper[4824]: E1209 10:17:06.726884 4824 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 10:17:06 crc kubenswrapper[4824]: E1209 10:17:06.727309 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r2pzz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-m6t59_openstack(8051e79f-a20f-4a10-b2f8-d8c770328797): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 10:17:06 crc kubenswrapper[4824]: E1209 10:17:06.729121 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-m6t59" podUID="8051e79f-a20f-4a10-b2f8-d8c770328797" Dec 09 10:17:06 crc kubenswrapper[4824]: I1209 10:17:06.770077 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55","Type":"ContainerDied","Data":"01b6e17da8db7c97663cfd93d2c005495e908868a882dc54077e7be6cff51a4f"} Dec 09 10:17:06 crc kubenswrapper[4824]: I1209 10:17:06.770128 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01b6e17da8db7c97663cfd93d2c005495e908868a882dc54077e7be6cff51a4f" Dec 09 10:17:06 crc kubenswrapper[4824]: I1209 10:17:06.926332 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.074937 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-erlang-cookie-secret\") pod \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.075662 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-config-data\") pod \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.075726 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-rabbitmq-tls\") pod \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.082446 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-rabbitmq-erlang-cookie\") pod \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.082715 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-pod-info\") pod \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.082832 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-rabbitmq-confd\") pod \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.082883 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-plugins-conf\") pod \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.083001 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-server-conf\") pod \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.101028 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" (UID: "09e7de63-6c82-4cb9-b9b6-12da1f3e6b55"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.103038 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" (UID: "09e7de63-6c82-4cb9-b9b6-12da1f3e6b55"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.360328 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c\") pod \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.360413 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-rabbitmq-plugins\") pod \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.369699 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" (UID: "09e7de63-6c82-4cb9-b9b6-12da1f3e6b55"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.370389 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-pod-info" (OuterVolumeSpecName: "pod-info") pod "09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" (UID: "09e7de63-6c82-4cb9-b9b6-12da1f3e6b55"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.370797 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" (UID: "09e7de63-6c82-4cb9-b9b6-12da1f3e6b55"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.374114 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-kube-api-access-p5pcr" (OuterVolumeSpecName: "kube-api-access-p5pcr") pod "09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" (UID: "09e7de63-6c82-4cb9-b9b6-12da1f3e6b55"). InnerVolumeSpecName "kube-api-access-p5pcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.377953 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5pcr\" (UniqueName: \"kubernetes.io/projected/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-kube-api-access-p5pcr\") pod \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\" (UID: \"09e7de63-6c82-4cb9-b9b6-12da1f3e6b55\") " Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.386395 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5pcr\" (UniqueName: \"kubernetes.io/projected/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-kube-api-access-p5pcr\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.387331 4824 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.387452 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.387537 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.387613 4824 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-pod-info\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.387701 4824 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.390042 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" (UID: "09e7de63-6c82-4cb9-b9b6-12da1f3e6b55"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.424343 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-config-data" (OuterVolumeSpecName: "config-data") pod "09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" (UID: "09e7de63-6c82-4cb9-b9b6-12da1f3e6b55"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.427427 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c" (OuterVolumeSpecName: "persistence") pod "09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" (UID: "09e7de63-6c82-4cb9-b9b6-12da1f3e6b55"). InnerVolumeSpecName "pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.490342 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c\") on node \"crc\" " Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.490378 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.490395 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.493183 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.592199 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/07c1822c-1257-4d8c-9a17-d921929000fa-server-conf\") pod \"07c1822c-1257-4d8c-9a17-d921929000fa\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.592335 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/07c1822c-1257-4d8c-9a17-d921929000fa-rabbitmq-plugins\") pod \"07c1822c-1257-4d8c-9a17-d921929000fa\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.592413 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/07c1822c-1257-4d8c-9a17-d921929000fa-plugins-conf\") pod \"07c1822c-1257-4d8c-9a17-d921929000fa\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.592448 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/07c1822c-1257-4d8c-9a17-d921929000fa-rabbitmq-confd\") pod \"07c1822c-1257-4d8c-9a17-d921929000fa\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.592544 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/07c1822c-1257-4d8c-9a17-d921929000fa-rabbitmq-erlang-cookie\") pod \"07c1822c-1257-4d8c-9a17-d921929000fa\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.592614 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/07c1822c-1257-4d8c-9a17-d921929000fa-config-data\") pod \"07c1822c-1257-4d8c-9a17-d921929000fa\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.592701 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/07c1822c-1257-4d8c-9a17-d921929000fa-rabbitmq-tls\") pod \"07c1822c-1257-4d8c-9a17-d921929000fa\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.592911 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/07c1822c-1257-4d8c-9a17-d921929000fa-pod-info\") pod \"07c1822c-1257-4d8c-9a17-d921929000fa\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.592951 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzd8c\" (UniqueName: \"kubernetes.io/projected/07c1822c-1257-4d8c-9a17-d921929000fa-kube-api-access-kzd8c\") pod \"07c1822c-1257-4d8c-9a17-d921929000fa\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.593119 4824 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.593316 4824 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c") on node "crc" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.594724 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07c1822c-1257-4d8c-9a17-d921929000fa-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "07c1822c-1257-4d8c-9a17-d921929000fa" (UID: "07c1822c-1257-4d8c-9a17-d921929000fa"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.594765 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07c1822c-1257-4d8c-9a17-d921929000fa-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "07c1822c-1257-4d8c-9a17-d921929000fa" (UID: "07c1822c-1257-4d8c-9a17-d921929000fa"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.594951 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07c1822c-1257-4d8c-9a17-d921929000fa-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "07c1822c-1257-4d8c-9a17-d921929000fa" (UID: "07c1822c-1257-4d8c-9a17-d921929000fa"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.595028 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3785e319-25c8-406d-9ae9-497d53624c0e\") pod \"07c1822c-1257-4d8c-9a17-d921929000fa\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.595534 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/07c1822c-1257-4d8c-9a17-d921929000fa-erlang-cookie-secret\") pod \"07c1822c-1257-4d8c-9a17-d921929000fa\" (UID: \"07c1822c-1257-4d8c-9a17-d921929000fa\") " Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.597495 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/07c1822c-1257-4d8c-9a17-d921929000fa-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.597533 4824 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/07c1822c-1257-4d8c-9a17-d921929000fa-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.597548 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/07c1822c-1257-4d8c-9a17-d921929000fa-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.597566 4824 reconciler_common.go:293] "Volume detached for volume \"pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.599549 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07c1822c-1257-4d8c-9a17-d921929000fa-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "07c1822c-1257-4d8c-9a17-d921929000fa" (UID: "07c1822c-1257-4d8c-9a17-d921929000fa"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.605150 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07c1822c-1257-4d8c-9a17-d921929000fa-kube-api-access-kzd8c" (OuterVolumeSpecName: "kube-api-access-kzd8c") pod "07c1822c-1257-4d8c-9a17-d921929000fa" (UID: "07c1822c-1257-4d8c-9a17-d921929000fa"). InnerVolumeSpecName "kube-api-access-kzd8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.605556 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-server-conf" (OuterVolumeSpecName: "server-conf") pod "09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" (UID: "09e7de63-6c82-4cb9-b9b6-12da1f3e6b55"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.605721 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/07c1822c-1257-4d8c-9a17-d921929000fa-pod-info" (OuterVolumeSpecName: "pod-info") pod "07c1822c-1257-4d8c-9a17-d921929000fa" (UID: "07c1822c-1257-4d8c-9a17-d921929000fa"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.605794 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07c1822c-1257-4d8c-9a17-d921929000fa-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "07c1822c-1257-4d8c-9a17-d921929000fa" (UID: "07c1822c-1257-4d8c-9a17-d921929000fa"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.628259 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07c1822c-1257-4d8c-9a17-d921929000fa-config-data" (OuterVolumeSpecName: "config-data") pod "07c1822c-1257-4d8c-9a17-d921929000fa" (UID: "07c1822c-1257-4d8c-9a17-d921929000fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.641045 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3785e319-25c8-406d-9ae9-497d53624c0e" (OuterVolumeSpecName: "persistence") pod "07c1822c-1257-4d8c-9a17-d921929000fa" (UID: "07c1822c-1257-4d8c-9a17-d921929000fa"). InnerVolumeSpecName "pvc-3785e319-25c8-406d-9ae9-497d53624c0e". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.652343 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" (UID: "09e7de63-6c82-4cb9-b9b6-12da1f3e6b55"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.699485 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07c1822c-1257-4d8c-9a17-d921929000fa-server-conf" (OuterVolumeSpecName: "server-conf") pod "07c1822c-1257-4d8c-9a17-d921929000fa" (UID: "07c1822c-1257-4d8c-9a17-d921929000fa"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.701225 4824 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-server-conf\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.701249 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/07c1822c-1257-4d8c-9a17-d921929000fa-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.701258 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/07c1822c-1257-4d8c-9a17-d921929000fa-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.701269 4824 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/07c1822c-1257-4d8c-9a17-d921929000fa-pod-info\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.701279 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzd8c\" (UniqueName: \"kubernetes.io/projected/07c1822c-1257-4d8c-9a17-d921929000fa-kube-api-access-kzd8c\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.701302 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-3785e319-25c8-406d-9ae9-497d53624c0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3785e319-25c8-406d-9ae9-497d53624c0e\") on node \"crc\" " Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.701314 4824 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/07c1822c-1257-4d8c-9a17-d921929000fa-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.701322 4824 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/07c1822c-1257-4d8c-9a17-d921929000fa-server-conf\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.701330 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.757185 4824 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.757758 4824 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-3785e319-25c8-406d-9ae9-497d53624c0e" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3785e319-25c8-406d-9ae9-497d53624c0e") on node "crc" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.789947 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"07c1822c-1257-4d8c-9a17-d921929000fa","Type":"ContainerDied","Data":"bc57fe0d8effd04b0a062b79287d0203b9abfd60d82ab15741c68514d9dab845"} Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.790010 4824 scope.go:117] "RemoveContainer" containerID="d35277030f4a65e4facfc048ac021fc3c08167f2f2aceae17375fa39c64384b6" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.790201 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.790428 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:07 crc kubenswrapper[4824]: E1209 10:17:07.794951 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-m6t59" podUID="8051e79f-a20f-4a10-b2f8-d8c770328797" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.803397 4824 reconciler_common.go:293] "Volume detached for volume \"pvc-3785e319-25c8-406d-9ae9-497d53624c0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3785e319-25c8-406d-9ae9-497d53624c0e\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.867058 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07c1822c-1257-4d8c-9a17-d921929000fa-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "07c1822c-1257-4d8c-9a17-d921929000fa" (UID: "07c1822c-1257-4d8c-9a17-d921929000fa"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.877422 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-x2dv9"] Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.897339 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 10:17:07 crc kubenswrapper[4824]: I1209 10:17:07.908524 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/07c1822c-1257-4d8c-9a17-d921929000fa-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.048147 4824 scope.go:117] "RemoveContainer" containerID="b631ae679dd053d822c043d7bb711544d02d155405d2eba54e7f52df4b151a2c" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.133057 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.160746 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.183689 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 10:17:08 crc kubenswrapper[4824]: E1209 10:17:08.185517 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07c1822c-1257-4d8c-9a17-d921929000fa" containerName="setup-container" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.185562 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="07c1822c-1257-4d8c-9a17-d921929000fa" containerName="setup-container" Dec 09 10:17:08 crc kubenswrapper[4824]: E1209 10:17:08.185602 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" containerName="setup-container" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.185611 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" containerName="setup-container" Dec 09 10:17:08 crc kubenswrapper[4824]: E1209 10:17:08.185655 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" containerName="rabbitmq" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.185664 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" containerName="rabbitmq" Dec 09 10:17:08 crc kubenswrapper[4824]: E1209 10:17:08.185689 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07c1822c-1257-4d8c-9a17-d921929000fa" containerName="rabbitmq" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.185701 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="07c1822c-1257-4d8c-9a17-d921929000fa" containerName="rabbitmq" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.186099 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="07c1822c-1257-4d8c-9a17-d921929000fa" containerName="rabbitmq" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.186116 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" containerName="rabbitmq" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.188026 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.190518 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-z2n6r" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.192148 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.192483 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.192654 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.194268 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.194483 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.195570 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.207260 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.320461 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b3de6262-0cdb-411a-9da1-9815c40d577e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.320517 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b3de6262-0cdb-411a-9da1-9815c40d577e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.320731 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b3de6262-0cdb-411a-9da1-9815c40d577e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.320937 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b3de6262-0cdb-411a-9da1-9815c40d577e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.321195 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdfff\" (UniqueName: \"kubernetes.io/projected/b3de6262-0cdb-411a-9da1-9815c40d577e-kube-api-access-qdfff\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.321327 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b3de6262-0cdb-411a-9da1-9815c40d577e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.321447 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3785e319-25c8-406d-9ae9-497d53624c0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3785e319-25c8-406d-9ae9-497d53624c0e\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.321498 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b3de6262-0cdb-411a-9da1-9815c40d577e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.321576 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b3de6262-0cdb-411a-9da1-9815c40d577e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.321640 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b3de6262-0cdb-411a-9da1-9815c40d577e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.321751 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b3de6262-0cdb-411a-9da1-9815c40d577e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.424563 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b3de6262-0cdb-411a-9da1-9815c40d577e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.424728 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b3de6262-0cdb-411a-9da1-9815c40d577e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.425088 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdfff\" (UniqueName: \"kubernetes.io/projected/b3de6262-0cdb-411a-9da1-9815c40d577e-kube-api-access-qdfff\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.425207 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b3de6262-0cdb-411a-9da1-9815c40d577e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.425251 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3785e319-25c8-406d-9ae9-497d53624c0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3785e319-25c8-406d-9ae9-497d53624c0e\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.425276 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b3de6262-0cdb-411a-9da1-9815c40d577e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.425304 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b3de6262-0cdb-411a-9da1-9815c40d577e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.425354 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b3de6262-0cdb-411a-9da1-9815c40d577e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.425419 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b3de6262-0cdb-411a-9da1-9815c40d577e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.425475 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b3de6262-0cdb-411a-9da1-9815c40d577e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.425502 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b3de6262-0cdb-411a-9da1-9815c40d577e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.425568 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b3de6262-0cdb-411a-9da1-9815c40d577e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.426141 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b3de6262-0cdb-411a-9da1-9815c40d577e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.426533 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b3de6262-0cdb-411a-9da1-9815c40d577e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.429090 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.430349 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3785e319-25c8-406d-9ae9-497d53624c0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3785e319-25c8-406d-9ae9-497d53624c0e\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/67b607a56a6ba5190d5a998ccfc91a99cdd5ecc60a8dca58e454bce5bec86124/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.430034 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b3de6262-0cdb-411a-9da1-9815c40d577e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.430289 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b3de6262-0cdb-411a-9da1-9815c40d577e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.430838 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b3de6262-0cdb-411a-9da1-9815c40d577e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.431385 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b3de6262-0cdb-411a-9da1-9815c40d577e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.433027 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b3de6262-0cdb-411a-9da1-9815c40d577e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.433201 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b3de6262-0cdb-411a-9da1-9815c40d577e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.443407 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdfff\" (UniqueName: \"kubernetes.io/projected/b3de6262-0cdb-411a-9da1-9815c40d577e-kube-api-access-qdfff\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.496557 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3785e319-25c8-406d-9ae9-497d53624c0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3785e319-25c8-406d-9ae9-497d53624c0e\") pod \"rabbitmq-cell1-server-0\" (UID: \"b3de6262-0cdb-411a-9da1-9815c40d577e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:08 crc kubenswrapper[4824]: I1209 10:17:08.577359 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:09 crc kubenswrapper[4824]: I1209 10:17:09.012573 4824 generic.go:334] "Generic (PLEG): container finished" podID="9ce40f74-fec2-4b15-bb84-d8972d806882" containerID="b523d6e0f8dafe752798ca4da4781efa6a4029524758a263a83bf9df93908f2e" exitCode=0 Dec 09 10:17:09 crc kubenswrapper[4824]: I1209 10:17:09.012913 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" event={"ID":"9ce40f74-fec2-4b15-bb84-d8972d806882","Type":"ContainerDied","Data":"b523d6e0f8dafe752798ca4da4781efa6a4029524758a263a83bf9df93908f2e"} Dec 09 10:17:09 crc kubenswrapper[4824]: I1209 10:17:09.012942 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" event={"ID":"9ce40f74-fec2-4b15-bb84-d8972d806882","Type":"ContainerStarted","Data":"edc7d6d649970d04f622de91c3228c671d21d70c87e2674d723ec0f9cb7283d4"} Dec 09 10:17:09 crc kubenswrapper[4824]: I1209 10:17:09.028014 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9339c10c-1cea-4ffc-8931-9b349a1e74cd","Type":"ContainerStarted","Data":"58b7bbcbca4a5299df56d0129fe8a9cbc58a3229e1cf89cdbd8c89a6391aa3f0"} Dec 09 10:17:09 crc kubenswrapper[4824]: I1209 10:17:09.417336 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 10:17:09 crc kubenswrapper[4824]: W1209 10:17:09.436327 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3de6262_0cdb_411a_9da1_9815c40d577e.slice/crio-c6dcd515c349b1a8612da45ffa5c805d6481c92eaf794bb0eb5287c9ad988ebb WatchSource:0}: Error finding container c6dcd515c349b1a8612da45ffa5c805d6481c92eaf794bb0eb5287c9ad988ebb: Status 404 returned error can't find the container with id c6dcd515c349b1a8612da45ffa5c805d6481c92eaf794bb0eb5287c9ad988ebb Dec 09 10:17:09 crc kubenswrapper[4824]: I1209 10:17:09.925539 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07c1822c-1257-4d8c-9a17-d921929000fa" path="/var/lib/kubelet/pods/07c1822c-1257-4d8c-9a17-d921929000fa/volumes" Dec 09 10:17:10 crc kubenswrapper[4824]: I1209 10:17:10.066071 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" event={"ID":"9ce40f74-fec2-4b15-bb84-d8972d806882","Type":"ContainerStarted","Data":"b22d28c5d6adc3f7dfe5eab3c3760c1d247b5fc9d91031e8b77ba664725bf202"} Dec 09 10:17:10 crc kubenswrapper[4824]: I1209 10:17:10.066138 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:10 crc kubenswrapper[4824]: I1209 10:17:10.067755 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b3de6262-0cdb-411a-9da1-9815c40d577e","Type":"ContainerStarted","Data":"c6dcd515c349b1a8612da45ffa5c805d6481c92eaf794bb0eb5287c9ad988ebb"} Dec 09 10:17:10 crc kubenswrapper[4824]: I1209 10:17:10.092761 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" podStartSLOduration=9.092738781 podStartE2EDuration="9.092738781s" podCreationTimestamp="2025-12-09 10:17:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:17:10.084911925 +0000 UTC m=+1786.419416592" watchObservedRunningTime="2025-12-09 10:17:10.092738781 +0000 UTC m=+1786.427243448" Dec 09 10:17:10 crc kubenswrapper[4824]: I1209 10:17:10.929670 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-2" podUID="09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.131:5671: i/o timeout" Dec 09 10:17:12 crc kubenswrapper[4824]: I1209 10:17:12.106159 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b3de6262-0cdb-411a-9da1-9815c40d577e","Type":"ContainerStarted","Data":"0f82d71db4d319456d8d62c49c7ec5a9bf02f3f31ca3986d9ac3eb0680d00b38"} Dec 09 10:17:12 crc kubenswrapper[4824]: I1209 10:17:12.131410 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="07c1822c-1257-4d8c-9a17-d921929000fa" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.132:5671: i/o timeout" Dec 09 10:17:13 crc kubenswrapper[4824]: I1209 10:17:13.318632 4824 scope.go:117] "RemoveContainer" containerID="17519c5dfbefe63576e185d8b281c4f8a3c5c8eaff5cd9a05b5db8801ca6f9b2" Dec 09 10:17:13 crc kubenswrapper[4824]: I1209 10:17:13.365209 4824 scope.go:117] "RemoveContainer" containerID="3ad239061116629b299b08f30e913aae04ed870beb2645bdf1ab9cfc59480f62" Dec 09 10:17:14 crc kubenswrapper[4824]: I1209 10:17:14.209935 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9339c10c-1cea-4ffc-8931-9b349a1e74cd","Type":"ContainerStarted","Data":"c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd"} Dec 09 10:17:15 crc kubenswrapper[4824]: I1209 10:17:15.223284 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9339c10c-1cea-4ffc-8931-9b349a1e74cd","Type":"ContainerStarted","Data":"3995b78d50fb1b44d9175e332201d4b380c4fffcac35d6bd8b74987d9f47009f"} Dec 09 10:17:16 crc kubenswrapper[4824]: I1209 10:17:16.238227 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9339c10c-1cea-4ffc-8931-9b349a1e74cd","Type":"ContainerStarted","Data":"bdb78b5ad1dc21e1979c54aa979963bb3d353abc40943ea4bbd889f390bb4342"} Dec 09 10:17:16 crc kubenswrapper[4824]: I1209 10:17:16.727038 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:16 crc kubenswrapper[4824]: I1209 10:17:16.801505 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn"] Dec 09 10:17:16 crc kubenswrapper[4824]: I1209 10:17:16.801851 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" podUID="a4e6339b-2aab-42c8-8783-5ffe08b2a026" containerName="dnsmasq-dns" containerID="cri-o://c1ac399e73a7f3af024137814b37c0b3bf0f8b0a0fbc29b162812fc959a015ce" gracePeriod=10 Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.002240 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-2vs9q"] Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.004710 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.020339 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-2vs9q"] Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.051184 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2510a71c-bf12-4207-96ee-6722c9d2febe-config\") pod \"dnsmasq-dns-6f6df4f56c-2vs9q\" (UID: \"2510a71c-bf12-4207-96ee-6722c9d2febe\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.051241 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2510a71c-bf12-4207-96ee-6722c9d2febe-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-2vs9q\" (UID: \"2510a71c-bf12-4207-96ee-6722c9d2febe\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.051263 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zgz7\" (UniqueName: \"kubernetes.io/projected/2510a71c-bf12-4207-96ee-6722c9d2febe-kube-api-access-6zgz7\") pod \"dnsmasq-dns-6f6df4f56c-2vs9q\" (UID: \"2510a71c-bf12-4207-96ee-6722c9d2febe\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.051396 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2510a71c-bf12-4207-96ee-6722c9d2febe-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-2vs9q\" (UID: \"2510a71c-bf12-4207-96ee-6722c9d2febe\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.051429 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2510a71c-bf12-4207-96ee-6722c9d2febe-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-2vs9q\" (UID: \"2510a71c-bf12-4207-96ee-6722c9d2febe\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.051450 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2510a71c-bf12-4207-96ee-6722c9d2febe-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-2vs9q\" (UID: \"2510a71c-bf12-4207-96ee-6722c9d2febe\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.051499 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2510a71c-bf12-4207-96ee-6722c9d2febe-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-2vs9q\" (UID: \"2510a71c-bf12-4207-96ee-6722c9d2febe\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.154909 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2510a71c-bf12-4207-96ee-6722c9d2febe-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-2vs9q\" (UID: \"2510a71c-bf12-4207-96ee-6722c9d2febe\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.154991 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2510a71c-bf12-4207-96ee-6722c9d2febe-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-2vs9q\" (UID: \"2510a71c-bf12-4207-96ee-6722c9d2febe\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.155022 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2510a71c-bf12-4207-96ee-6722c9d2febe-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-2vs9q\" (UID: \"2510a71c-bf12-4207-96ee-6722c9d2febe\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.155087 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2510a71c-bf12-4207-96ee-6722c9d2febe-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-2vs9q\" (UID: \"2510a71c-bf12-4207-96ee-6722c9d2febe\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.155227 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2510a71c-bf12-4207-96ee-6722c9d2febe-config\") pod \"dnsmasq-dns-6f6df4f56c-2vs9q\" (UID: \"2510a71c-bf12-4207-96ee-6722c9d2febe\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.155263 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2510a71c-bf12-4207-96ee-6722c9d2febe-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-2vs9q\" (UID: \"2510a71c-bf12-4207-96ee-6722c9d2febe\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.155290 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zgz7\" (UniqueName: \"kubernetes.io/projected/2510a71c-bf12-4207-96ee-6722c9d2febe-kube-api-access-6zgz7\") pod \"dnsmasq-dns-6f6df4f56c-2vs9q\" (UID: \"2510a71c-bf12-4207-96ee-6722c9d2febe\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.156573 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2510a71c-bf12-4207-96ee-6722c9d2febe-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-2vs9q\" (UID: \"2510a71c-bf12-4207-96ee-6722c9d2febe\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.156616 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2510a71c-bf12-4207-96ee-6722c9d2febe-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-2vs9q\" (UID: \"2510a71c-bf12-4207-96ee-6722c9d2febe\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.214067 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2510a71c-bf12-4207-96ee-6722c9d2febe-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-2vs9q\" (UID: \"2510a71c-bf12-4207-96ee-6722c9d2febe\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.220296 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2510a71c-bf12-4207-96ee-6722c9d2febe-config\") pod \"dnsmasq-dns-6f6df4f56c-2vs9q\" (UID: \"2510a71c-bf12-4207-96ee-6722c9d2febe\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.220316 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2510a71c-bf12-4207-96ee-6722c9d2febe-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-2vs9q\" (UID: \"2510a71c-bf12-4207-96ee-6722c9d2febe\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.220607 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2510a71c-bf12-4207-96ee-6722c9d2febe-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-2vs9q\" (UID: \"2510a71c-bf12-4207-96ee-6722c9d2febe\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.259914 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zgz7\" (UniqueName: \"kubernetes.io/projected/2510a71c-bf12-4207-96ee-6722c9d2febe-kube-api-access-6zgz7\") pod \"dnsmasq-dns-6f6df4f56c-2vs9q\" (UID: \"2510a71c-bf12-4207-96ee-6722c9d2febe\") " pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.288099 4824 generic.go:334] "Generic (PLEG): container finished" podID="a4e6339b-2aab-42c8-8783-5ffe08b2a026" containerID="c1ac399e73a7f3af024137814b37c0b3bf0f8b0a0fbc29b162812fc959a015ce" exitCode=0 Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.288153 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" event={"ID":"a4e6339b-2aab-42c8-8783-5ffe08b2a026","Type":"ContainerDied","Data":"c1ac399e73a7f3af024137814b37c0b3bf0f8b0a0fbc29b162812fc959a015ce"} Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.337587 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.717091 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.778352 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-dns-svc\") pod \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\" (UID: \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\") " Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.778604 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnv56\" (UniqueName: \"kubernetes.io/projected/a4e6339b-2aab-42c8-8783-5ffe08b2a026-kube-api-access-cnv56\") pod \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\" (UID: \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\") " Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.778674 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-dns-swift-storage-0\") pod \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\" (UID: \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\") " Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.778764 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-config\") pod \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\" (UID: \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\") " Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.778927 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-ovsdbserver-nb\") pod \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\" (UID: \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\") " Dec 09 10:17:17 crc kubenswrapper[4824]: I1209 10:17:17.778960 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-ovsdbserver-sb\") pod \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\" (UID: \"a4e6339b-2aab-42c8-8783-5ffe08b2a026\") " Dec 09 10:17:18 crc kubenswrapper[4824]: I1209 10:17:18.196081 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a4e6339b-2aab-42c8-8783-5ffe08b2a026" (UID: "a4e6339b-2aab-42c8-8783-5ffe08b2a026"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:17:18 crc kubenswrapper[4824]: I1209 10:17:18.278992 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a4e6339b-2aab-42c8-8783-5ffe08b2a026" (UID: "a4e6339b-2aab-42c8-8783-5ffe08b2a026"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:17:18 crc kubenswrapper[4824]: I1209 10:17:18.284527 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a4e6339b-2aab-42c8-8783-5ffe08b2a026" (UID: "a4e6339b-2aab-42c8-8783-5ffe08b2a026"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:17:18 crc kubenswrapper[4824]: W1209 10:17:18.298219 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2510a71c_bf12_4207_96ee_6722c9d2febe.slice/crio-5679aaaf4e94074da60199e245416d1e96097e2addcc6b9d17417d91187ba44e WatchSource:0}: Error finding container 5679aaaf4e94074da60199e245416d1e96097e2addcc6b9d17417d91187ba44e: Status 404 returned error can't find the container with id 5679aaaf4e94074da60199e245416d1e96097e2addcc6b9d17417d91187ba44e Dec 09 10:17:18 crc kubenswrapper[4824]: I1209 10:17:18.299433 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:18 crc kubenswrapper[4824]: I1209 10:17:18.299470 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:18 crc kubenswrapper[4824]: I1209 10:17:18.299485 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:18 crc kubenswrapper[4824]: I1209 10:17:18.301041 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4e6339b-2aab-42c8-8783-5ffe08b2a026-kube-api-access-cnv56" (OuterVolumeSpecName: "kube-api-access-cnv56") pod "a4e6339b-2aab-42c8-8783-5ffe08b2a026" (UID: "a4e6339b-2aab-42c8-8783-5ffe08b2a026"). InnerVolumeSpecName "kube-api-access-cnv56". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:17:18 crc kubenswrapper[4824]: I1209 10:17:18.321990 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" Dec 09 10:17:18 crc kubenswrapper[4824]: I1209 10:17:18.339275 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-2vs9q"] Dec 09 10:17:18 crc kubenswrapper[4824]: I1209 10:17:18.339327 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" event={"ID":"a4e6339b-2aab-42c8-8783-5ffe08b2a026","Type":"ContainerDied","Data":"f208b3a0de338f20acefd3d43724d58c6b46c52eb5854ab0a041516276a7e8c8"} Dec 09 10:17:18 crc kubenswrapper[4824]: I1209 10:17:18.339372 4824 scope.go:117] "RemoveContainer" containerID="c1ac399e73a7f3af024137814b37c0b3bf0f8b0a0fbc29b162812fc959a015ce" Dec 09 10:17:18 crc kubenswrapper[4824]: I1209 10:17:18.342383 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9339c10c-1cea-4ffc-8931-9b349a1e74cd","Type":"ContainerStarted","Data":"2ce9839174d941dad5d87c2e7b97f38166a4af17a4b596061df35d667bd94f17"} Dec 09 10:17:18 crc kubenswrapper[4824]: I1209 10:17:18.342660 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 10:17:18 crc kubenswrapper[4824]: I1209 10:17:18.375903 4824 scope.go:117] "RemoveContainer" containerID="4cc3c3aebaf7c99d9342eb91ffcdacf5d4611de118217c6bc4ec4df04d37ce0b" Dec 09 10:17:18 crc kubenswrapper[4824]: I1209 10:17:18.392334 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=11.848364992 podStartE2EDuration="21.392304598s" podCreationTimestamp="2025-12-09 10:16:57 +0000 UTC" firstStartedPulling="2025-12-09 10:17:07.883000627 +0000 UTC m=+1784.217505294" lastFinishedPulling="2025-12-09 10:17:17.426940233 +0000 UTC m=+1793.761444900" observedRunningTime="2025-12-09 10:17:18.377842293 +0000 UTC m=+1794.712346980" watchObservedRunningTime="2025-12-09 10:17:18.392304598 +0000 UTC m=+1794.726809265" Dec 09 10:17:18 crc kubenswrapper[4824]: I1209 10:17:18.401421 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnv56\" (UniqueName: \"kubernetes.io/projected/a4e6339b-2aab-42c8-8783-5ffe08b2a026-kube-api-access-cnv56\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:18 crc kubenswrapper[4824]: I1209 10:17:18.405212 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a4e6339b-2aab-42c8-8783-5ffe08b2a026" (UID: "a4e6339b-2aab-42c8-8783-5ffe08b2a026"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:17:18 crc kubenswrapper[4824]: I1209 10:17:18.423710 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-config" (OuterVolumeSpecName: "config") pod "a4e6339b-2aab-42c8-8783-5ffe08b2a026" (UID: "a4e6339b-2aab-42c8-8783-5ffe08b2a026"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:17:18 crc kubenswrapper[4824]: I1209 10:17:18.507806 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:18 crc kubenswrapper[4824]: I1209 10:17:18.507835 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a4e6339b-2aab-42c8-8783-5ffe08b2a026-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:18 crc kubenswrapper[4824]: I1209 10:17:18.716678 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn"] Dec 09 10:17:18 crc kubenswrapper[4824]: I1209 10:17:18.729355 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn"] Dec 09 10:17:19 crc kubenswrapper[4824]: I1209 10:17:19.043643 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:17:19 crc kubenswrapper[4824]: E1209 10:17:19.043899 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:17:19 crc kubenswrapper[4824]: I1209 10:17:19.364176 4824 generic.go:334] "Generic (PLEG): container finished" podID="2510a71c-bf12-4207-96ee-6722c9d2febe" containerID="12035a6a1ff5ea944bff6f8dee3e7e1704c05a0ec9f7111aded246b5506cf588" exitCode=0 Dec 09 10:17:19 crc kubenswrapper[4824]: I1209 10:17:19.366523 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" event={"ID":"2510a71c-bf12-4207-96ee-6722c9d2febe","Type":"ContainerDied","Data":"12035a6a1ff5ea944bff6f8dee3e7e1704c05a0ec9f7111aded246b5506cf588"} Dec 09 10:17:19 crc kubenswrapper[4824]: I1209 10:17:19.366560 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" event={"ID":"2510a71c-bf12-4207-96ee-6722c9d2febe","Type":"ContainerStarted","Data":"5679aaaf4e94074da60199e245416d1e96097e2addcc6b9d17417d91187ba44e"} Dec 09 10:17:20 crc kubenswrapper[4824]: I1209 10:17:20.112773 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4e6339b-2aab-42c8-8783-5ffe08b2a026" path="/var/lib/kubelet/pods/a4e6339b-2aab-42c8-8783-5ffe08b2a026/volumes" Dec 09 10:17:20 crc kubenswrapper[4824]: I1209 10:17:20.380010 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-m6t59" event={"ID":"8051e79f-a20f-4a10-b2f8-d8c770328797","Type":"ContainerStarted","Data":"d0ab6558e5bc53fd8b72cf22f8e4de08649e93342f3afbc2f9b46f91abd831ba"} Dec 09 10:17:20 crc kubenswrapper[4824]: I1209 10:17:20.385040 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" event={"ID":"2510a71c-bf12-4207-96ee-6722c9d2febe","Type":"ContainerStarted","Data":"3ccc50ba7bc4958d291995f31fc224161afbd0d2c29e51842990200d730b8d0e"} Dec 09 10:17:20 crc kubenswrapper[4824]: I1209 10:17:20.385511 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:20 crc kubenswrapper[4824]: I1209 10:17:20.398444 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-m6t59" podStartSLOduration=2.723211424 podStartE2EDuration="39.398423995s" podCreationTimestamp="2025-12-09 10:16:41 +0000 UTC" firstStartedPulling="2025-12-09 10:16:42.63161617 +0000 UTC m=+1758.966120827" lastFinishedPulling="2025-12-09 10:17:19.306828731 +0000 UTC m=+1795.641333398" observedRunningTime="2025-12-09 10:17:20.397249588 +0000 UTC m=+1796.731754265" watchObservedRunningTime="2025-12-09 10:17:20.398423995 +0000 UTC m=+1796.732928662" Dec 09 10:17:20 crc kubenswrapper[4824]: I1209 10:17:20.418747 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" podStartSLOduration=4.418727695 podStartE2EDuration="4.418727695s" podCreationTimestamp="2025-12-09 10:17:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:17:20.416425342 +0000 UTC m=+1796.750930019" watchObservedRunningTime="2025-12-09 10:17:20.418727695 +0000 UTC m=+1796.753232362" Dec 09 10:17:22 crc kubenswrapper[4824]: I1209 10:17:22.261265 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6b7bbf7cf9-lhfdn" podUID="a4e6339b-2aab-42c8-8783-5ffe08b2a026" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.0:5353: i/o timeout" Dec 09 10:17:23 crc kubenswrapper[4824]: I1209 10:17:23.462461 4824 generic.go:334] "Generic (PLEG): container finished" podID="8051e79f-a20f-4a10-b2f8-d8c770328797" containerID="d0ab6558e5bc53fd8b72cf22f8e4de08649e93342f3afbc2f9b46f91abd831ba" exitCode=0 Dec 09 10:17:23 crc kubenswrapper[4824]: I1209 10:17:23.462558 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-m6t59" event={"ID":"8051e79f-a20f-4a10-b2f8-d8c770328797","Type":"ContainerDied","Data":"d0ab6558e5bc53fd8b72cf22f8e4de08649e93342f3afbc2f9b46f91abd831ba"} Dec 09 10:17:24 crc kubenswrapper[4824]: I1209 10:17:24.909597 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-m6t59" Dec 09 10:17:25 crc kubenswrapper[4824]: I1209 10:17:25.008936 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2pzz\" (UniqueName: \"kubernetes.io/projected/8051e79f-a20f-4a10-b2f8-d8c770328797-kube-api-access-r2pzz\") pod \"8051e79f-a20f-4a10-b2f8-d8c770328797\" (UID: \"8051e79f-a20f-4a10-b2f8-d8c770328797\") " Dec 09 10:17:25 crc kubenswrapper[4824]: I1209 10:17:25.009270 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8051e79f-a20f-4a10-b2f8-d8c770328797-config-data\") pod \"8051e79f-a20f-4a10-b2f8-d8c770328797\" (UID: \"8051e79f-a20f-4a10-b2f8-d8c770328797\") " Dec 09 10:17:25 crc kubenswrapper[4824]: I1209 10:17:25.009341 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8051e79f-a20f-4a10-b2f8-d8c770328797-combined-ca-bundle\") pod \"8051e79f-a20f-4a10-b2f8-d8c770328797\" (UID: \"8051e79f-a20f-4a10-b2f8-d8c770328797\") " Dec 09 10:17:25 crc kubenswrapper[4824]: I1209 10:17:25.015250 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8051e79f-a20f-4a10-b2f8-d8c770328797-kube-api-access-r2pzz" (OuterVolumeSpecName: "kube-api-access-r2pzz") pod "8051e79f-a20f-4a10-b2f8-d8c770328797" (UID: "8051e79f-a20f-4a10-b2f8-d8c770328797"). InnerVolumeSpecName "kube-api-access-r2pzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:17:25 crc kubenswrapper[4824]: I1209 10:17:25.067660 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8051e79f-a20f-4a10-b2f8-d8c770328797-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8051e79f-a20f-4a10-b2f8-d8c770328797" (UID: "8051e79f-a20f-4a10-b2f8-d8c770328797"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:17:25 crc kubenswrapper[4824]: I1209 10:17:25.117590 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2pzz\" (UniqueName: \"kubernetes.io/projected/8051e79f-a20f-4a10-b2f8-d8c770328797-kube-api-access-r2pzz\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:25 crc kubenswrapper[4824]: I1209 10:17:25.117639 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8051e79f-a20f-4a10-b2f8-d8c770328797-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:25 crc kubenswrapper[4824]: I1209 10:17:25.128940 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8051e79f-a20f-4a10-b2f8-d8c770328797-config-data" (OuterVolumeSpecName: "config-data") pod "8051e79f-a20f-4a10-b2f8-d8c770328797" (UID: "8051e79f-a20f-4a10-b2f8-d8c770328797"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:17:25 crc kubenswrapper[4824]: I1209 10:17:25.219387 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8051e79f-a20f-4a10-b2f8-d8c770328797-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:25 crc kubenswrapper[4824]: I1209 10:17:25.489497 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-m6t59" event={"ID":"8051e79f-a20f-4a10-b2f8-d8c770328797","Type":"ContainerDied","Data":"3b3174fe8553a469c25ed20745542d8c596a60c10b553d1cabcba892ef1c9060"} Dec 09 10:17:25 crc kubenswrapper[4824]: I1209 10:17:25.489539 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b3174fe8553a469c25ed20745542d8c596a60c10b553d1cabcba892ef1c9060" Dec 09 10:17:25 crc kubenswrapper[4824]: I1209 10:17:25.489636 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-m6t59" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.391249 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-67ffc9d677-6k2l9"] Dec 09 10:17:26 crc kubenswrapper[4824]: E1209 10:17:26.392107 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8051e79f-a20f-4a10-b2f8-d8c770328797" containerName="heat-db-sync" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.392121 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8051e79f-a20f-4a10-b2f8-d8c770328797" containerName="heat-db-sync" Dec 09 10:17:26 crc kubenswrapper[4824]: E1209 10:17:26.392150 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4e6339b-2aab-42c8-8783-5ffe08b2a026" containerName="init" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.392156 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4e6339b-2aab-42c8-8783-5ffe08b2a026" containerName="init" Dec 09 10:17:26 crc kubenswrapper[4824]: E1209 10:17:26.392174 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4e6339b-2aab-42c8-8783-5ffe08b2a026" containerName="dnsmasq-dns" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.392181 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4e6339b-2aab-42c8-8783-5ffe08b2a026" containerName="dnsmasq-dns" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.392418 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8051e79f-a20f-4a10-b2f8-d8c770328797" containerName="heat-db-sync" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.392449 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4e6339b-2aab-42c8-8783-5ffe08b2a026" containerName="dnsmasq-dns" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.393309 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-67ffc9d677-6k2l9" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.406651 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-67ffc9d677-6k2l9"] Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.451024 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-69c99fc8cd-j8cz5"] Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.453247 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.467351 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-57cb8dfdcd-c62t6"] Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.469941 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-57cb8dfdcd-c62t6" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.490403 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-69c99fc8cd-j8cz5"] Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.510817 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-57cb8dfdcd-c62t6"] Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.551441 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce017433-0624-4b3f-b1e9-0de3b591a602-combined-ca-bundle\") pod \"heat-engine-67ffc9d677-6k2l9\" (UID: \"ce017433-0624-4b3f-b1e9-0de3b591a602\") " pod="openstack/heat-engine-67ffc9d677-6k2l9" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.551540 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtl42\" (UniqueName: \"kubernetes.io/projected/ce017433-0624-4b3f-b1e9-0de3b591a602-kube-api-access-xtl42\") pod \"heat-engine-67ffc9d677-6k2l9\" (UID: \"ce017433-0624-4b3f-b1e9-0de3b591a602\") " pod="openstack/heat-engine-67ffc9d677-6k2l9" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.551593 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ec002ca-0e91-48ab-a982-51d6c19b33b6-config-data-custom\") pod \"heat-cfnapi-69c99fc8cd-j8cz5\" (UID: \"1ec002ca-0e91-48ab-a982-51d6c19b33b6\") " pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.551627 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ec002ca-0e91-48ab-a982-51d6c19b33b6-public-tls-certs\") pod \"heat-cfnapi-69c99fc8cd-j8cz5\" (UID: \"1ec002ca-0e91-48ab-a982-51d6c19b33b6\") " pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.551664 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cr2qv\" (UniqueName: \"kubernetes.io/projected/1ec002ca-0e91-48ab-a982-51d6c19b33b6-kube-api-access-cr2qv\") pod \"heat-cfnapi-69c99fc8cd-j8cz5\" (UID: \"1ec002ca-0e91-48ab-a982-51d6c19b33b6\") " pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.551694 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ce017433-0624-4b3f-b1e9-0de3b591a602-config-data-custom\") pod \"heat-engine-67ffc9d677-6k2l9\" (UID: \"ce017433-0624-4b3f-b1e9-0de3b591a602\") " pod="openstack/heat-engine-67ffc9d677-6k2l9" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.551731 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ec002ca-0e91-48ab-a982-51d6c19b33b6-combined-ca-bundle\") pod \"heat-cfnapi-69c99fc8cd-j8cz5\" (UID: \"1ec002ca-0e91-48ab-a982-51d6c19b33b6\") " pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.551823 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ec002ca-0e91-48ab-a982-51d6c19b33b6-config-data\") pod \"heat-cfnapi-69c99fc8cd-j8cz5\" (UID: \"1ec002ca-0e91-48ab-a982-51d6c19b33b6\") " pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.551862 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce017433-0624-4b3f-b1e9-0de3b591a602-config-data\") pod \"heat-engine-67ffc9d677-6k2l9\" (UID: \"ce017433-0624-4b3f-b1e9-0de3b591a602\") " pod="openstack/heat-engine-67ffc9d677-6k2l9" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.551906 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ec002ca-0e91-48ab-a982-51d6c19b33b6-internal-tls-certs\") pod \"heat-cfnapi-69c99fc8cd-j8cz5\" (UID: \"1ec002ca-0e91-48ab-a982-51d6c19b33b6\") " pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.653533 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce017433-0624-4b3f-b1e9-0de3b591a602-config-data\") pod \"heat-engine-67ffc9d677-6k2l9\" (UID: \"ce017433-0624-4b3f-b1e9-0de3b591a602\") " pod="openstack/heat-engine-67ffc9d677-6k2l9" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.653620 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ec002ca-0e91-48ab-a982-51d6c19b33b6-internal-tls-certs\") pod \"heat-cfnapi-69c99fc8cd-j8cz5\" (UID: \"1ec002ca-0e91-48ab-a982-51d6c19b33b6\") " pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.653657 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6dnm\" (UniqueName: \"kubernetes.io/projected/aee6f3d2-13da-489e-844c-70355d34d102-kube-api-access-g6dnm\") pod \"heat-api-57cb8dfdcd-c62t6\" (UID: \"aee6f3d2-13da-489e-844c-70355d34d102\") " pod="openstack/heat-api-57cb8dfdcd-c62t6" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.653723 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aee6f3d2-13da-489e-844c-70355d34d102-internal-tls-certs\") pod \"heat-api-57cb8dfdcd-c62t6\" (UID: \"aee6f3d2-13da-489e-844c-70355d34d102\") " pod="openstack/heat-api-57cb8dfdcd-c62t6" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.653759 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce017433-0624-4b3f-b1e9-0de3b591a602-combined-ca-bundle\") pod \"heat-engine-67ffc9d677-6k2l9\" (UID: \"ce017433-0624-4b3f-b1e9-0de3b591a602\") " pod="openstack/heat-engine-67ffc9d677-6k2l9" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.653838 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aee6f3d2-13da-489e-844c-70355d34d102-public-tls-certs\") pod \"heat-api-57cb8dfdcd-c62t6\" (UID: \"aee6f3d2-13da-489e-844c-70355d34d102\") " pod="openstack/heat-api-57cb8dfdcd-c62t6" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.653872 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aee6f3d2-13da-489e-844c-70355d34d102-combined-ca-bundle\") pod \"heat-api-57cb8dfdcd-c62t6\" (UID: \"aee6f3d2-13da-489e-844c-70355d34d102\") " pod="openstack/heat-api-57cb8dfdcd-c62t6" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.653925 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtl42\" (UniqueName: \"kubernetes.io/projected/ce017433-0624-4b3f-b1e9-0de3b591a602-kube-api-access-xtl42\") pod \"heat-engine-67ffc9d677-6k2l9\" (UID: \"ce017433-0624-4b3f-b1e9-0de3b591a602\") " pod="openstack/heat-engine-67ffc9d677-6k2l9" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.653977 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ec002ca-0e91-48ab-a982-51d6c19b33b6-config-data-custom\") pod \"heat-cfnapi-69c99fc8cd-j8cz5\" (UID: \"1ec002ca-0e91-48ab-a982-51d6c19b33b6\") " pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.654010 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ec002ca-0e91-48ab-a982-51d6c19b33b6-public-tls-certs\") pod \"heat-cfnapi-69c99fc8cd-j8cz5\" (UID: \"1ec002ca-0e91-48ab-a982-51d6c19b33b6\") " pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.654044 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cr2qv\" (UniqueName: \"kubernetes.io/projected/1ec002ca-0e91-48ab-a982-51d6c19b33b6-kube-api-access-cr2qv\") pod \"heat-cfnapi-69c99fc8cd-j8cz5\" (UID: \"1ec002ca-0e91-48ab-a982-51d6c19b33b6\") " pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.654065 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ce017433-0624-4b3f-b1e9-0de3b591a602-config-data-custom\") pod \"heat-engine-67ffc9d677-6k2l9\" (UID: \"ce017433-0624-4b3f-b1e9-0de3b591a602\") " pod="openstack/heat-engine-67ffc9d677-6k2l9" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.654134 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ec002ca-0e91-48ab-a982-51d6c19b33b6-combined-ca-bundle\") pod \"heat-cfnapi-69c99fc8cd-j8cz5\" (UID: \"1ec002ca-0e91-48ab-a982-51d6c19b33b6\") " pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.654173 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aee6f3d2-13da-489e-844c-70355d34d102-config-data-custom\") pod \"heat-api-57cb8dfdcd-c62t6\" (UID: \"aee6f3d2-13da-489e-844c-70355d34d102\") " pod="openstack/heat-api-57cb8dfdcd-c62t6" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.654196 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aee6f3d2-13da-489e-844c-70355d34d102-config-data\") pod \"heat-api-57cb8dfdcd-c62t6\" (UID: \"aee6f3d2-13da-489e-844c-70355d34d102\") " pod="openstack/heat-api-57cb8dfdcd-c62t6" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.654258 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ec002ca-0e91-48ab-a982-51d6c19b33b6-config-data\") pod \"heat-cfnapi-69c99fc8cd-j8cz5\" (UID: \"1ec002ca-0e91-48ab-a982-51d6c19b33b6\") " pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.659584 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ec002ca-0e91-48ab-a982-51d6c19b33b6-config-data\") pod \"heat-cfnapi-69c99fc8cd-j8cz5\" (UID: \"1ec002ca-0e91-48ab-a982-51d6c19b33b6\") " pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.659742 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ec002ca-0e91-48ab-a982-51d6c19b33b6-combined-ca-bundle\") pod \"heat-cfnapi-69c99fc8cd-j8cz5\" (UID: \"1ec002ca-0e91-48ab-a982-51d6c19b33b6\") " pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.659752 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ec002ca-0e91-48ab-a982-51d6c19b33b6-internal-tls-certs\") pod \"heat-cfnapi-69c99fc8cd-j8cz5\" (UID: \"1ec002ca-0e91-48ab-a982-51d6c19b33b6\") " pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.661350 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ec002ca-0e91-48ab-a982-51d6c19b33b6-config-data-custom\") pod \"heat-cfnapi-69c99fc8cd-j8cz5\" (UID: \"1ec002ca-0e91-48ab-a982-51d6c19b33b6\") " pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.661383 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ec002ca-0e91-48ab-a982-51d6c19b33b6-public-tls-certs\") pod \"heat-cfnapi-69c99fc8cd-j8cz5\" (UID: \"1ec002ca-0e91-48ab-a982-51d6c19b33b6\") " pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.661723 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce017433-0624-4b3f-b1e9-0de3b591a602-config-data\") pod \"heat-engine-67ffc9d677-6k2l9\" (UID: \"ce017433-0624-4b3f-b1e9-0de3b591a602\") " pod="openstack/heat-engine-67ffc9d677-6k2l9" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.661882 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce017433-0624-4b3f-b1e9-0de3b591a602-combined-ca-bundle\") pod \"heat-engine-67ffc9d677-6k2l9\" (UID: \"ce017433-0624-4b3f-b1e9-0de3b591a602\") " pod="openstack/heat-engine-67ffc9d677-6k2l9" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.668818 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ce017433-0624-4b3f-b1e9-0de3b591a602-config-data-custom\") pod \"heat-engine-67ffc9d677-6k2l9\" (UID: \"ce017433-0624-4b3f-b1e9-0de3b591a602\") " pod="openstack/heat-engine-67ffc9d677-6k2l9" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.672387 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cr2qv\" (UniqueName: \"kubernetes.io/projected/1ec002ca-0e91-48ab-a982-51d6c19b33b6-kube-api-access-cr2qv\") pod \"heat-cfnapi-69c99fc8cd-j8cz5\" (UID: \"1ec002ca-0e91-48ab-a982-51d6c19b33b6\") " pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.684468 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtl42\" (UniqueName: \"kubernetes.io/projected/ce017433-0624-4b3f-b1e9-0de3b591a602-kube-api-access-xtl42\") pod \"heat-engine-67ffc9d677-6k2l9\" (UID: \"ce017433-0624-4b3f-b1e9-0de3b591a602\") " pod="openstack/heat-engine-67ffc9d677-6k2l9" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.722916 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-67ffc9d677-6k2l9" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.761239 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aee6f3d2-13da-489e-844c-70355d34d102-config-data-custom\") pod \"heat-api-57cb8dfdcd-c62t6\" (UID: \"aee6f3d2-13da-489e-844c-70355d34d102\") " pod="openstack/heat-api-57cb8dfdcd-c62t6" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.761302 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aee6f3d2-13da-489e-844c-70355d34d102-config-data\") pod \"heat-api-57cb8dfdcd-c62t6\" (UID: \"aee6f3d2-13da-489e-844c-70355d34d102\") " pod="openstack/heat-api-57cb8dfdcd-c62t6" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.761455 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6dnm\" (UniqueName: \"kubernetes.io/projected/aee6f3d2-13da-489e-844c-70355d34d102-kube-api-access-g6dnm\") pod \"heat-api-57cb8dfdcd-c62t6\" (UID: \"aee6f3d2-13da-489e-844c-70355d34d102\") " pod="openstack/heat-api-57cb8dfdcd-c62t6" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.761529 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aee6f3d2-13da-489e-844c-70355d34d102-internal-tls-certs\") pod \"heat-api-57cb8dfdcd-c62t6\" (UID: \"aee6f3d2-13da-489e-844c-70355d34d102\") " pod="openstack/heat-api-57cb8dfdcd-c62t6" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.761597 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aee6f3d2-13da-489e-844c-70355d34d102-public-tls-certs\") pod \"heat-api-57cb8dfdcd-c62t6\" (UID: \"aee6f3d2-13da-489e-844c-70355d34d102\") " pod="openstack/heat-api-57cb8dfdcd-c62t6" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.761630 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aee6f3d2-13da-489e-844c-70355d34d102-combined-ca-bundle\") pod \"heat-api-57cb8dfdcd-c62t6\" (UID: \"aee6f3d2-13da-489e-844c-70355d34d102\") " pod="openstack/heat-api-57cb8dfdcd-c62t6" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.766713 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aee6f3d2-13da-489e-844c-70355d34d102-config-data-custom\") pod \"heat-api-57cb8dfdcd-c62t6\" (UID: \"aee6f3d2-13da-489e-844c-70355d34d102\") " pod="openstack/heat-api-57cb8dfdcd-c62t6" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.767967 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aee6f3d2-13da-489e-844c-70355d34d102-config-data\") pod \"heat-api-57cb8dfdcd-c62t6\" (UID: \"aee6f3d2-13da-489e-844c-70355d34d102\") " pod="openstack/heat-api-57cb8dfdcd-c62t6" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.769872 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aee6f3d2-13da-489e-844c-70355d34d102-public-tls-certs\") pod \"heat-api-57cb8dfdcd-c62t6\" (UID: \"aee6f3d2-13da-489e-844c-70355d34d102\") " pod="openstack/heat-api-57cb8dfdcd-c62t6" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.769933 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aee6f3d2-13da-489e-844c-70355d34d102-internal-tls-certs\") pod \"heat-api-57cb8dfdcd-c62t6\" (UID: \"aee6f3d2-13da-489e-844c-70355d34d102\") " pod="openstack/heat-api-57cb8dfdcd-c62t6" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.770133 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aee6f3d2-13da-489e-844c-70355d34d102-combined-ca-bundle\") pod \"heat-api-57cb8dfdcd-c62t6\" (UID: \"aee6f3d2-13da-489e-844c-70355d34d102\") " pod="openstack/heat-api-57cb8dfdcd-c62t6" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.785568 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6dnm\" (UniqueName: \"kubernetes.io/projected/aee6f3d2-13da-489e-844c-70355d34d102-kube-api-access-g6dnm\") pod \"heat-api-57cb8dfdcd-c62t6\" (UID: \"aee6f3d2-13da-489e-844c-70355d34d102\") " pod="openstack/heat-api-57cb8dfdcd-c62t6" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.804447 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" Dec 09 10:17:26 crc kubenswrapper[4824]: I1209 10:17:26.811661 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-57cb8dfdcd-c62t6" Dec 09 10:17:27 crc kubenswrapper[4824]: W1209 10:17:27.303368 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce017433_0624_4b3f_b1e9_0de3b591a602.slice/crio-34edea345068e1cba4a6e984c2d9ec2b67305859c53a75a05666c0973cdc371d WatchSource:0}: Error finding container 34edea345068e1cba4a6e984c2d9ec2b67305859c53a75a05666c0973cdc371d: Status 404 returned error can't find the container with id 34edea345068e1cba4a6e984c2d9ec2b67305859c53a75a05666c0973cdc371d Dec 09 10:17:27 crc kubenswrapper[4824]: I1209 10:17:27.310734 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-67ffc9d677-6k2l9"] Dec 09 10:17:27 crc kubenswrapper[4824]: I1209 10:17:27.339470 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6f6df4f56c-2vs9q" Dec 09 10:17:27 crc kubenswrapper[4824]: I1209 10:17:27.441545 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-x2dv9"] Dec 09 10:17:27 crc kubenswrapper[4824]: I1209 10:17:27.441922 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" podUID="9ce40f74-fec2-4b15-bb84-d8972d806882" containerName="dnsmasq-dns" containerID="cri-o://b22d28c5d6adc3f7dfe5eab3c3760c1d247b5fc9d91031e8b77ba664725bf202" gracePeriod=10 Dec 09 10:17:27 crc kubenswrapper[4824]: I1209 10:17:27.486530 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-69c99fc8cd-j8cz5"] Dec 09 10:17:27 crc kubenswrapper[4824]: I1209 10:17:27.526723 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-67ffc9d677-6k2l9" event={"ID":"ce017433-0624-4b3f-b1e9-0de3b591a602","Type":"ContainerStarted","Data":"34edea345068e1cba4a6e984c2d9ec2b67305859c53a75a05666c0973cdc371d"} Dec 09 10:17:27 crc kubenswrapper[4824]: I1209 10:17:27.530823 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-57cb8dfdcd-c62t6"] Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.109221 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.116183 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-dns-svc\") pod \"9ce40f74-fec2-4b15-bb84-d8972d806882\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.116495 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqmf9\" (UniqueName: \"kubernetes.io/projected/9ce40f74-fec2-4b15-bb84-d8972d806882-kube-api-access-tqmf9\") pod \"9ce40f74-fec2-4b15-bb84-d8972d806882\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.116532 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-ovsdbserver-nb\") pod \"9ce40f74-fec2-4b15-bb84-d8972d806882\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.116747 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-config\") pod \"9ce40f74-fec2-4b15-bb84-d8972d806882\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.116820 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-dns-swift-storage-0\") pod \"9ce40f74-fec2-4b15-bb84-d8972d806882\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.116918 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-openstack-edpm-ipam\") pod \"9ce40f74-fec2-4b15-bb84-d8972d806882\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.117022 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-ovsdbserver-sb\") pod \"9ce40f74-fec2-4b15-bb84-d8972d806882\" (UID: \"9ce40f74-fec2-4b15-bb84-d8972d806882\") " Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.122317 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ce40f74-fec2-4b15-bb84-d8972d806882-kube-api-access-tqmf9" (OuterVolumeSpecName: "kube-api-access-tqmf9") pod "9ce40f74-fec2-4b15-bb84-d8972d806882" (UID: "9ce40f74-fec2-4b15-bb84-d8972d806882"). InnerVolumeSpecName "kube-api-access-tqmf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.218093 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9ce40f74-fec2-4b15-bb84-d8972d806882" (UID: "9ce40f74-fec2-4b15-bb84-d8972d806882"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.219881 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqmf9\" (UniqueName: \"kubernetes.io/projected/9ce40f74-fec2-4b15-bb84-d8972d806882-kube-api-access-tqmf9\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.219902 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.239172 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9ce40f74-fec2-4b15-bb84-d8972d806882" (UID: "9ce40f74-fec2-4b15-bb84-d8972d806882"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.264229 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-config" (OuterVolumeSpecName: "config") pod "9ce40f74-fec2-4b15-bb84-d8972d806882" (UID: "9ce40f74-fec2-4b15-bb84-d8972d806882"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.277977 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9ce40f74-fec2-4b15-bb84-d8972d806882" (UID: "9ce40f74-fec2-4b15-bb84-d8972d806882"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.292217 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "9ce40f74-fec2-4b15-bb84-d8972d806882" (UID: "9ce40f74-fec2-4b15-bb84-d8972d806882"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.321471 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9ce40f74-fec2-4b15-bb84-d8972d806882" (UID: "9ce40f74-fec2-4b15-bb84-d8972d806882"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.322808 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.322843 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-config\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.322855 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.322865 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.322874 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ce40f74-fec2-4b15-bb84-d8972d806882-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.547530 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" event={"ID":"1ec002ca-0e91-48ab-a982-51d6c19b33b6","Type":"ContainerStarted","Data":"5251dde272ff973a1b6b069e4aad92eb5c9446724f120b6de28ce1a59f7cae1a"} Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.550515 4824 generic.go:334] "Generic (PLEG): container finished" podID="9ce40f74-fec2-4b15-bb84-d8972d806882" containerID="b22d28c5d6adc3f7dfe5eab3c3760c1d247b5fc9d91031e8b77ba664725bf202" exitCode=0 Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.550567 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" event={"ID":"9ce40f74-fec2-4b15-bb84-d8972d806882","Type":"ContainerDied","Data":"b22d28c5d6adc3f7dfe5eab3c3760c1d247b5fc9d91031e8b77ba664725bf202"} Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.550603 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.550638 4824 scope.go:117] "RemoveContainer" containerID="b22d28c5d6adc3f7dfe5eab3c3760c1d247b5fc9d91031e8b77ba664725bf202" Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.550623 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-x2dv9" event={"ID":"9ce40f74-fec2-4b15-bb84-d8972d806882","Type":"ContainerDied","Data":"edc7d6d649970d04f622de91c3228c671d21d70c87e2674d723ec0f9cb7283d4"} Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.554758 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-57cb8dfdcd-c62t6" event={"ID":"aee6f3d2-13da-489e-844c-70355d34d102","Type":"ContainerStarted","Data":"8c02bd0fb7e0b310f23ded33573af5f98154260b452fd54412f28713a153c0f1"} Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.557081 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-67ffc9d677-6k2l9" event={"ID":"ce017433-0624-4b3f-b1e9-0de3b591a602","Type":"ContainerStarted","Data":"4df2c8df0f678bd1ccf62f4fc2dc3aaf928fae10f8c8a84b71de286a3d6cb2d1"} Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.557315 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-67ffc9d677-6k2l9" Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.590498 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-67ffc9d677-6k2l9" podStartSLOduration=2.590475763 podStartE2EDuration="2.590475763s" podCreationTimestamp="2025-12-09 10:17:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:17:28.580268831 +0000 UTC m=+1804.914773498" watchObservedRunningTime="2025-12-09 10:17:28.590475763 +0000 UTC m=+1804.924980430" Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.611020 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-x2dv9"] Dec 09 10:17:28 crc kubenswrapper[4824]: I1209 10:17:28.627346 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-x2dv9"] Dec 09 10:17:29 crc kubenswrapper[4824]: I1209 10:17:29.346363 4824 scope.go:117] "RemoveContainer" containerID="b523d6e0f8dafe752798ca4da4781efa6a4029524758a263a83bf9df93908f2e" Dec 09 10:17:29 crc kubenswrapper[4824]: I1209 10:17:29.383152 4824 scope.go:117] "RemoveContainer" containerID="b22d28c5d6adc3f7dfe5eab3c3760c1d247b5fc9d91031e8b77ba664725bf202" Dec 09 10:17:29 crc kubenswrapper[4824]: E1209 10:17:29.383717 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b22d28c5d6adc3f7dfe5eab3c3760c1d247b5fc9d91031e8b77ba664725bf202\": container with ID starting with b22d28c5d6adc3f7dfe5eab3c3760c1d247b5fc9d91031e8b77ba664725bf202 not found: ID does not exist" containerID="b22d28c5d6adc3f7dfe5eab3c3760c1d247b5fc9d91031e8b77ba664725bf202" Dec 09 10:17:29 crc kubenswrapper[4824]: I1209 10:17:29.383761 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b22d28c5d6adc3f7dfe5eab3c3760c1d247b5fc9d91031e8b77ba664725bf202"} err="failed to get container status \"b22d28c5d6adc3f7dfe5eab3c3760c1d247b5fc9d91031e8b77ba664725bf202\": rpc error: code = NotFound desc = could not find container \"b22d28c5d6adc3f7dfe5eab3c3760c1d247b5fc9d91031e8b77ba664725bf202\": container with ID starting with b22d28c5d6adc3f7dfe5eab3c3760c1d247b5fc9d91031e8b77ba664725bf202 not found: ID does not exist" Dec 09 10:17:29 crc kubenswrapper[4824]: I1209 10:17:29.383806 4824 scope.go:117] "RemoveContainer" containerID="b523d6e0f8dafe752798ca4da4781efa6a4029524758a263a83bf9df93908f2e" Dec 09 10:17:29 crc kubenswrapper[4824]: E1209 10:17:29.384067 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b523d6e0f8dafe752798ca4da4781efa6a4029524758a263a83bf9df93908f2e\": container with ID starting with b523d6e0f8dafe752798ca4da4781efa6a4029524758a263a83bf9df93908f2e not found: ID does not exist" containerID="b523d6e0f8dafe752798ca4da4781efa6a4029524758a263a83bf9df93908f2e" Dec 09 10:17:29 crc kubenswrapper[4824]: I1209 10:17:29.384097 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b523d6e0f8dafe752798ca4da4781efa6a4029524758a263a83bf9df93908f2e"} err="failed to get container status \"b523d6e0f8dafe752798ca4da4781efa6a4029524758a263a83bf9df93908f2e\": rpc error: code = NotFound desc = could not find container \"b523d6e0f8dafe752798ca4da4781efa6a4029524758a263a83bf9df93908f2e\": container with ID starting with b523d6e0f8dafe752798ca4da4781efa6a4029524758a263a83bf9df93908f2e not found: ID does not exist" Dec 09 10:17:29 crc kubenswrapper[4824]: I1209 10:17:29.931071 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ce40f74-fec2-4b15-bb84-d8972d806882" path="/var/lib/kubelet/pods/9ce40f74-fec2-4b15-bb84-d8972d806882/volumes" Dec 09 10:17:30 crc kubenswrapper[4824]: I1209 10:17:30.587453 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" event={"ID":"1ec002ca-0e91-48ab-a982-51d6c19b33b6","Type":"ContainerStarted","Data":"b818d84cf1f27c5d870d1036d52d38d49d136bfcc38fbd426e99a500ef776041"} Dec 09 10:17:30 crc kubenswrapper[4824]: I1209 10:17:30.587775 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" Dec 09 10:17:30 crc kubenswrapper[4824]: I1209 10:17:30.589608 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-57cb8dfdcd-c62t6" event={"ID":"aee6f3d2-13da-489e-844c-70355d34d102","Type":"ContainerStarted","Data":"36d282364ea777fb482b3a1295bac5e2a3b65e05fbfd13e2d72a0f4288546f16"} Dec 09 10:17:30 crc kubenswrapper[4824]: I1209 10:17:30.589737 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-57cb8dfdcd-c62t6" Dec 09 10:17:30 crc kubenswrapper[4824]: I1209 10:17:30.604311 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" podStartSLOduration=2.794828443 podStartE2EDuration="4.604293442s" podCreationTimestamp="2025-12-09 10:17:26 +0000 UTC" firstStartedPulling="2025-12-09 10:17:27.544391494 +0000 UTC m=+1803.878896161" lastFinishedPulling="2025-12-09 10:17:29.353856493 +0000 UTC m=+1805.688361160" observedRunningTime="2025-12-09 10:17:30.60293893 +0000 UTC m=+1806.937443617" watchObservedRunningTime="2025-12-09 10:17:30.604293442 +0000 UTC m=+1806.938798099" Dec 09 10:17:30 crc kubenswrapper[4824]: I1209 10:17:30.639425 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-57cb8dfdcd-c62t6" podStartSLOduration=2.820475412 podStartE2EDuration="4.639402039s" podCreationTimestamp="2025-12-09 10:17:26 +0000 UTC" firstStartedPulling="2025-12-09 10:17:27.542079441 +0000 UTC m=+1803.876584108" lastFinishedPulling="2025-12-09 10:17:29.361006058 +0000 UTC m=+1805.695510735" observedRunningTime="2025-12-09 10:17:30.623851059 +0000 UTC m=+1806.958355746" watchObservedRunningTime="2025-12-09 10:17:30.639402039 +0000 UTC m=+1806.973906706" Dec 09 10:17:33 crc kubenswrapper[4824]: I1209 10:17:33.919754 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:17:33 crc kubenswrapper[4824]: E1209 10:17:33.920586 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:17:37 crc kubenswrapper[4824]: I1209 10:17:37.892353 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j"] Dec 09 10:17:37 crc kubenswrapper[4824]: E1209 10:17:37.893506 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ce40f74-fec2-4b15-bb84-d8972d806882" containerName="dnsmasq-dns" Dec 09 10:17:37 crc kubenswrapper[4824]: I1209 10:17:37.893526 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ce40f74-fec2-4b15-bb84-d8972d806882" containerName="dnsmasq-dns" Dec 09 10:17:37 crc kubenswrapper[4824]: E1209 10:17:37.893548 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ce40f74-fec2-4b15-bb84-d8972d806882" containerName="init" Dec 09 10:17:37 crc kubenswrapper[4824]: I1209 10:17:37.893556 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ce40f74-fec2-4b15-bb84-d8972d806882" containerName="init" Dec 09 10:17:37 crc kubenswrapper[4824]: I1209 10:17:37.893923 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ce40f74-fec2-4b15-bb84-d8972d806882" containerName="dnsmasq-dns" Dec 09 10:17:37 crc kubenswrapper[4824]: I1209 10:17:37.895276 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j" Dec 09 10:17:37 crc kubenswrapper[4824]: I1209 10:17:37.900420 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 10:17:37 crc kubenswrapper[4824]: I1209 10:17:37.900711 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 10:17:37 crc kubenswrapper[4824]: I1209 10:17:37.900900 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 10:17:37 crc kubenswrapper[4824]: I1209 10:17:37.901057 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5nzld" Dec 09 10:17:37 crc kubenswrapper[4824]: I1209 10:17:37.937439 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j"] Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.022557 4824 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","burstable","pod09e7de63-6c82-4cb9-b9b6-12da1f3e6b55"] err="unable to destroy cgroup paths for cgroup [kubepods burstable pod09e7de63-6c82-4cb9-b9b6-12da1f3e6b55] : Timed out while waiting for systemd to remove kubepods-burstable-pod09e7de63_6c82_4cb9_b9b6_12da1f3e6b55.slice" Dec 09 10:17:38 crc kubenswrapper[4824]: E1209 10:17:38.022609 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods burstable pod09e7de63-6c82-4cb9-b9b6-12da1f3e6b55] : unable to destroy cgroup paths for cgroup [kubepods burstable pod09e7de63-6c82-4cb9-b9b6-12da1f3e6b55] : Timed out while waiting for systemd to remove kubepods-burstable-pod09e7de63_6c82_4cb9_b9b6_12da1f3e6b55.slice" pod="openstack/rabbitmq-server-2" podUID="09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.057667 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e1db872-7d04-4c89-b725-9e3d84fed0fb-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j\" (UID: \"6e1db872-7d04-4c89-b725-9e3d84fed0fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j" Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.057729 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp4fh\" (UniqueName: \"kubernetes.io/projected/6e1db872-7d04-4c89-b725-9e3d84fed0fb-kube-api-access-tp4fh\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j\" (UID: \"6e1db872-7d04-4c89-b725-9e3d84fed0fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j" Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.057806 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e1db872-7d04-4c89-b725-9e3d84fed0fb-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j\" (UID: \"6e1db872-7d04-4c89-b725-9e3d84fed0fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j" Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.058661 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e1db872-7d04-4c89-b725-9e3d84fed0fb-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j\" (UID: \"6e1db872-7d04-4c89-b725-9e3d84fed0fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j" Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.162130 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e1db872-7d04-4c89-b725-9e3d84fed0fb-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j\" (UID: \"6e1db872-7d04-4c89-b725-9e3d84fed0fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j" Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.162428 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e1db872-7d04-4c89-b725-9e3d84fed0fb-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j\" (UID: \"6e1db872-7d04-4c89-b725-9e3d84fed0fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j" Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.162556 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e1db872-7d04-4c89-b725-9e3d84fed0fb-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j\" (UID: \"6e1db872-7d04-4c89-b725-9e3d84fed0fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j" Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.162605 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp4fh\" (UniqueName: \"kubernetes.io/projected/6e1db872-7d04-4c89-b725-9e3d84fed0fb-kube-api-access-tp4fh\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j\" (UID: \"6e1db872-7d04-4c89-b725-9e3d84fed0fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j" Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.169658 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e1db872-7d04-4c89-b725-9e3d84fed0fb-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j\" (UID: \"6e1db872-7d04-4c89-b725-9e3d84fed0fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j" Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.170077 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e1db872-7d04-4c89-b725-9e3d84fed0fb-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j\" (UID: \"6e1db872-7d04-4c89-b725-9e3d84fed0fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j" Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.171194 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e1db872-7d04-4c89-b725-9e3d84fed0fb-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j\" (UID: \"6e1db872-7d04-4c89-b725-9e3d84fed0fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j" Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.180700 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp4fh\" (UniqueName: \"kubernetes.io/projected/6e1db872-7d04-4c89-b725-9e3d84fed0fb-kube-api-access-tp4fh\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j\" (UID: \"6e1db872-7d04-4c89-b725-9e3d84fed0fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j" Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.229464 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j" Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.445160 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-57cb8dfdcd-c62t6" Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.484611 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-69c99fc8cd-j8cz5" Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.616306 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-98f88b69-9xf7z"] Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.616559 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-98f88b69-9xf7z" podUID="d6e9dedb-7873-4008-a9a0-fa90ec7c99f1" containerName="heat-api" containerID="cri-o://1004fb0fd749d85593aaa49a6c0f575039b7bf13fb8560d1104508bdc20a30fe" gracePeriod=60 Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.634192 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-ff546578f-d6scd"] Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.634413 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-ff546578f-d6scd" podUID="14358ed8-0c10-48f7-a1b1-37e75e7a6400" containerName="heat-cfnapi" containerID="cri-o://4d69b95b0579a8d3ad4ea33d1c22f7fdffb90d75e0a488beff69658977be69de" gracePeriod=60 Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.679158 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.771083 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-2"] Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.825567 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-2"] Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.847446 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-2"] Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.850192 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.866555 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-2"] Dec 09 10:17:38 crc kubenswrapper[4824]: I1209 10:17:38.999398 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c97a0c28-3a5a-430a-8000-e5c0518262b3-config-data\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:38.999641 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c97a0c28-3a5a-430a-8000-e5c0518262b3-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:38.999846 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c97a0c28-3a5a-430a-8000-e5c0518262b3-pod-info\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.000245 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7lnp\" (UniqueName: \"kubernetes.io/projected/c97a0c28-3a5a-430a-8000-e5c0518262b3-kube-api-access-v7lnp\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.000297 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c97a0c28-3a5a-430a-8000-e5c0518262b3-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.000365 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c97a0c28-3a5a-430a-8000-e5c0518262b3-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.000451 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c97a0c28-3a5a-430a-8000-e5c0518262b3-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.000541 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c97a0c28-3a5a-430a-8000-e5c0518262b3-server-conf\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.000582 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.000613 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c97a0c28-3a5a-430a-8000-e5c0518262b3-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.000661 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c97a0c28-3a5a-430a-8000-e5c0518262b3-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.089363 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j"] Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.093082 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.102735 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7lnp\" (UniqueName: \"kubernetes.io/projected/c97a0c28-3a5a-430a-8000-e5c0518262b3-kube-api-access-v7lnp\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.102848 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c97a0c28-3a5a-430a-8000-e5c0518262b3-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.102900 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c97a0c28-3a5a-430a-8000-e5c0518262b3-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.102963 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c97a0c28-3a5a-430a-8000-e5c0518262b3-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.103028 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c97a0c28-3a5a-430a-8000-e5c0518262b3-server-conf\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.103054 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.103092 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c97a0c28-3a5a-430a-8000-e5c0518262b3-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.103127 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c97a0c28-3a5a-430a-8000-e5c0518262b3-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.103236 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c97a0c28-3a5a-430a-8000-e5c0518262b3-config-data\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.103343 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c97a0c28-3a5a-430a-8000-e5c0518262b3-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.103430 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c97a0c28-3a5a-430a-8000-e5c0518262b3-pod-info\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.104862 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c97a0c28-3a5a-430a-8000-e5c0518262b3-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.105330 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c97a0c28-3a5a-430a-8000-e5c0518262b3-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.108080 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c97a0c28-3a5a-430a-8000-e5c0518262b3-config-data\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.110717 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c97a0c28-3a5a-430a-8000-e5c0518262b3-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.111238 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c97a0c28-3a5a-430a-8000-e5c0518262b3-server-conf\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.112276 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c97a0c28-3a5a-430a-8000-e5c0518262b3-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.115358 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.117324 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4a25c77492aa5c7e92a555e527d30ff343f34e4f3d4bb5f823cc7b33af361462/globalmount\"" pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.119468 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c97a0c28-3a5a-430a-8000-e5c0518262b3-pod-info\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.125913 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c97a0c28-3a5a-430a-8000-e5c0518262b3-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.126484 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c97a0c28-3a5a-430a-8000-e5c0518262b3-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.127533 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7lnp\" (UniqueName: \"kubernetes.io/projected/c97a0c28-3a5a-430a-8000-e5c0518262b3-kube-api-access-v7lnp\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.197481 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-775b9e78-8921-43e4-a9b5-17afed3cf89c\") pod \"rabbitmq-server-2\" (UID: \"c97a0c28-3a5a-430a-8000-e5c0518262b3\") " pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.496630 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.711250 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j" event={"ID":"6e1db872-7d04-4c89-b725-9e3d84fed0fb","Type":"ContainerStarted","Data":"32431b06114800a867d9bd7c8974545de76e793a5c57384fffe2f03d5a602396"} Dec 09 10:17:39 crc kubenswrapper[4824]: I1209 10:17:39.944217 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09e7de63-6c82-4cb9-b9b6-12da1f3e6b55" path="/var/lib/kubelet/pods/09e7de63-6c82-4cb9-b9b6-12da1f3e6b55/volumes" Dec 09 10:17:40 crc kubenswrapper[4824]: I1209 10:17:40.002918 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-2"] Dec 09 10:17:40 crc kubenswrapper[4824]: W1209 10:17:40.009712 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc97a0c28_3a5a_430a_8000_e5c0518262b3.slice/crio-80619f927485fe000a3549dfb4c7f7727b20b087021b3a3a27705e5b1c11b01f WatchSource:0}: Error finding container 80619f927485fe000a3549dfb4c7f7727b20b087021b3a3a27705e5b1c11b01f: Status 404 returned error can't find the container with id 80619f927485fe000a3549dfb4c7f7727b20b087021b3a3a27705e5b1c11b01f Dec 09 10:17:40 crc kubenswrapper[4824]: I1209 10:17:40.725216 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"c97a0c28-3a5a-430a-8000-e5c0518262b3","Type":"ContainerStarted","Data":"80619f927485fe000a3549dfb4c7f7727b20b087021b3a3a27705e5b1c11b01f"} Dec 09 10:17:41 crc kubenswrapper[4824]: I1209 10:17:41.875389 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-ff546578f-d6scd" podUID="14358ed8-0c10-48f7-a1b1-37e75e7a6400" containerName="heat-cfnapi" probeResult="failure" output="Get \"https://10.217.0.219:8000/healthcheck\": read tcp 10.217.0.2:59248->10.217.0.219:8000: read: connection reset by peer" Dec 09 10:17:41 crc kubenswrapper[4824]: I1209 10:17:41.919166 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-98f88b69-9xf7z" podUID="d6e9dedb-7873-4008-a9a0-fa90ec7c99f1" containerName="heat-api" probeResult="failure" output="Get \"https://10.217.0.218:8004/healthcheck\": read tcp 10.217.0.2:36474->10.217.0.218:8004: read: connection reset by peer" Dec 09 10:17:42 crc kubenswrapper[4824]: I1209 10:17:42.755042 4824 generic.go:334] "Generic (PLEG): container finished" podID="d6e9dedb-7873-4008-a9a0-fa90ec7c99f1" containerID="1004fb0fd749d85593aaa49a6c0f575039b7bf13fb8560d1104508bdc20a30fe" exitCode=0 Dec 09 10:17:42 crc kubenswrapper[4824]: I1209 10:17:42.755247 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-98f88b69-9xf7z" event={"ID":"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1","Type":"ContainerDied","Data":"1004fb0fd749d85593aaa49a6c0f575039b7bf13fb8560d1104508bdc20a30fe"} Dec 09 10:17:42 crc kubenswrapper[4824]: I1209 10:17:42.757998 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"c97a0c28-3a5a-430a-8000-e5c0518262b3","Type":"ContainerStarted","Data":"50c050207a10cb427846e70c07ba22db4dfae172195a3de54910a2b8b534fb61"} Dec 09 10:17:42 crc kubenswrapper[4824]: I1209 10:17:42.799415 4824 generic.go:334] "Generic (PLEG): container finished" podID="14358ed8-0c10-48f7-a1b1-37e75e7a6400" containerID="4d69b95b0579a8d3ad4ea33d1c22f7fdffb90d75e0a488beff69658977be69de" exitCode=0 Dec 09 10:17:42 crc kubenswrapper[4824]: I1209 10:17:42.799466 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-ff546578f-d6scd" event={"ID":"14358ed8-0c10-48f7-a1b1-37e75e7a6400","Type":"ContainerDied","Data":"4d69b95b0579a8d3ad4ea33d1c22f7fdffb90d75e0a488beff69658977be69de"} Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.420740 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-98f88b69-9xf7z" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.447160 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8lpt\" (UniqueName: \"kubernetes.io/projected/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-kube-api-access-f8lpt\") pod \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\" (UID: \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\") " Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.447304 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-internal-tls-certs\") pod \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\" (UID: \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\") " Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.447334 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-config-data\") pod \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\" (UID: \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\") " Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.447526 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-public-tls-certs\") pod \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\" (UID: \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\") " Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.447564 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-config-data-custom\") pod \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\" (UID: \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\") " Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.447633 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-combined-ca-bundle\") pod \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\" (UID: \"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1\") " Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.477814 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d6e9dedb-7873-4008-a9a0-fa90ec7c99f1" (UID: "d6e9dedb-7873-4008-a9a0-fa90ec7c99f1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.483112 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-kube-api-access-f8lpt" (OuterVolumeSpecName: "kube-api-access-f8lpt") pod "d6e9dedb-7873-4008-a9a0-fa90ec7c99f1" (UID: "d6e9dedb-7873-4008-a9a0-fa90ec7c99f1"). InnerVolumeSpecName "kube-api-access-f8lpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.535266 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6e9dedb-7873-4008-a9a0-fa90ec7c99f1" (UID: "d6e9dedb-7873-4008-a9a0-fa90ec7c99f1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.549672 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-ff546578f-d6scd" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.550056 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.550080 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.550089 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8lpt\" (UniqueName: \"kubernetes.io/projected/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-kube-api-access-f8lpt\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.570114 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d6e9dedb-7873-4008-a9a0-fa90ec7c99f1" (UID: "d6e9dedb-7873-4008-a9a0-fa90ec7c99f1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.571025 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-config-data" (OuterVolumeSpecName: "config-data") pod "d6e9dedb-7873-4008-a9a0-fa90ec7c99f1" (UID: "d6e9dedb-7873-4008-a9a0-fa90ec7c99f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.578908 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d6e9dedb-7873-4008-a9a0-fa90ec7c99f1" (UID: "d6e9dedb-7873-4008-a9a0-fa90ec7c99f1"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.651411 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-internal-tls-certs\") pod \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\" (UID: \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\") " Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.651582 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-config-data-custom\") pod \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\" (UID: \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\") " Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.651632 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-combined-ca-bundle\") pod \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\" (UID: \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\") " Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.651746 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-public-tls-certs\") pod \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\" (UID: \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\") " Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.651813 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-config-data\") pod \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\" (UID: \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\") " Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.651830 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6skhn\" (UniqueName: \"kubernetes.io/projected/14358ed8-0c10-48f7-a1b1-37e75e7a6400-kube-api-access-6skhn\") pod \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\" (UID: \"14358ed8-0c10-48f7-a1b1-37e75e7a6400\") " Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.652630 4824 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.652655 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.652665 4824 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.656270 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "14358ed8-0c10-48f7-a1b1-37e75e7a6400" (UID: "14358ed8-0c10-48f7-a1b1-37e75e7a6400"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.658388 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14358ed8-0c10-48f7-a1b1-37e75e7a6400-kube-api-access-6skhn" (OuterVolumeSpecName: "kube-api-access-6skhn") pod "14358ed8-0c10-48f7-a1b1-37e75e7a6400" (UID: "14358ed8-0c10-48f7-a1b1-37e75e7a6400"). InnerVolumeSpecName "kube-api-access-6skhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.706661 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14358ed8-0c10-48f7-a1b1-37e75e7a6400" (UID: "14358ed8-0c10-48f7-a1b1-37e75e7a6400"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.722487 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "14358ed8-0c10-48f7-a1b1-37e75e7a6400" (UID: "14358ed8-0c10-48f7-a1b1-37e75e7a6400"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.723050 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "14358ed8-0c10-48f7-a1b1-37e75e7a6400" (UID: "14358ed8-0c10-48f7-a1b1-37e75e7a6400"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.756192 4824 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.756241 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.756254 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.756267 4824 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.756284 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6skhn\" (UniqueName: \"kubernetes.io/projected/14358ed8-0c10-48f7-a1b1-37e75e7a6400-kube-api-access-6skhn\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.758032 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-config-data" (OuterVolumeSpecName: "config-data") pod "14358ed8-0c10-48f7-a1b1-37e75e7a6400" (UID: "14358ed8-0c10-48f7-a1b1-37e75e7a6400"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.816156 4824 generic.go:334] "Generic (PLEG): container finished" podID="b3de6262-0cdb-411a-9da1-9815c40d577e" containerID="0f82d71db4d319456d8d62c49c7ec5a9bf02f3f31ca3986d9ac3eb0680d00b38" exitCode=0 Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.816249 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b3de6262-0cdb-411a-9da1-9815c40d577e","Type":"ContainerDied","Data":"0f82d71db4d319456d8d62c49c7ec5a9bf02f3f31ca3986d9ac3eb0680d00b38"} Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.819822 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-98f88b69-9xf7z" event={"ID":"d6e9dedb-7873-4008-a9a0-fa90ec7c99f1","Type":"ContainerDied","Data":"6d34a1e4bfc2eaeebf01f7268e2d032e85b89e81f2b3d82e9ccb1d7a05f51a28"} Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.819849 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-98f88b69-9xf7z" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.819868 4824 scope.go:117] "RemoveContainer" containerID="1004fb0fd749d85593aaa49a6c0f575039b7bf13fb8560d1104508bdc20a30fe" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.821631 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-ff546578f-d6scd" event={"ID":"14358ed8-0c10-48f7-a1b1-37e75e7a6400","Type":"ContainerDied","Data":"7c4b1e2e5473b7a828801198d6ff9a886d4e523f0f32395ce793221796eb0132"} Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.821658 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-ff546578f-d6scd" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.858572 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14358ed8-0c10-48f7-a1b1-37e75e7a6400-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.876268 4824 scope.go:117] "RemoveContainer" containerID="4d69b95b0579a8d3ad4ea33d1c22f7fdffb90d75e0a488beff69658977be69de" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.888892 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-ff546578f-d6scd"] Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.906980 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-ff546578f-d6scd"] Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.926846 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14358ed8-0c10-48f7-a1b1-37e75e7a6400" path="/var/lib/kubelet/pods/14358ed8-0c10-48f7-a1b1-37e75e7a6400/volumes" Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.927626 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-98f88b69-9xf7z"] Dec 09 10:17:43 crc kubenswrapper[4824]: I1209 10:17:43.943744 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-98f88b69-9xf7z"] Dec 09 10:17:44 crc kubenswrapper[4824]: I1209 10:17:44.837345 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b3de6262-0cdb-411a-9da1-9815c40d577e","Type":"ContainerStarted","Data":"7a654d829e3a6d1e734e4d984369d09f4b7b54676bb2879603598fc9b996f224"} Dec 09 10:17:44 crc kubenswrapper[4824]: I1209 10:17:44.837996 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:44 crc kubenswrapper[4824]: I1209 10:17:44.869217 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.869193829 podStartE2EDuration="36.869193829s" podCreationTimestamp="2025-12-09 10:17:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:17:44.863644114 +0000 UTC m=+1821.198148781" watchObservedRunningTime="2025-12-09 10:17:44.869193829 +0000 UTC m=+1821.203698496" Dec 09 10:17:45 crc kubenswrapper[4824]: I1209 10:17:45.931207 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6e9dedb-7873-4008-a9a0-fa90ec7c99f1" path="/var/lib/kubelet/pods/d6e9dedb-7873-4008-a9a0-fa90ec7c99f1/volumes" Dec 09 10:17:46 crc kubenswrapper[4824]: I1209 10:17:46.772929 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-67ffc9d677-6k2l9" Dec 09 10:17:46 crc kubenswrapper[4824]: I1209 10:17:46.828135 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-7d54cc949d-rzfx4"] Dec 09 10:17:46 crc kubenswrapper[4824]: I1209 10:17:46.828372 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-7d54cc949d-rzfx4" podUID="d8154a7d-1dbc-4fa3-a862-140f2fd807b2" containerName="heat-engine" containerID="cri-o://6a711400741bf3c06a5572c6a46ac91b8ec31fb5fdfda32e199126f0b080b249" gracePeriod=60 Dec 09 10:17:48 crc kubenswrapper[4824]: I1209 10:17:48.911323 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:17:48 crc kubenswrapper[4824]: E1209 10:17:48.912095 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:17:51 crc kubenswrapper[4824]: I1209 10:17:51.896410 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-jvpcc"] Dec 09 10:17:51 crc kubenswrapper[4824]: I1209 10:17:51.934510 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-jvpcc"] Dec 09 10:17:51 crc kubenswrapper[4824]: E1209 10:17:51.967266 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6a711400741bf3c06a5572c6a46ac91b8ec31fb5fdfda32e199126f0b080b249" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 09 10:17:51 crc kubenswrapper[4824]: E1209 10:17:51.971723 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6a711400741bf3c06a5572c6a46ac91b8ec31fb5fdfda32e199126f0b080b249" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 09 10:17:51 crc kubenswrapper[4824]: E1209 10:17:51.982850 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6a711400741bf3c06a5572c6a46ac91b8ec31fb5fdfda32e199126f0b080b249" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 09 10:17:51 crc kubenswrapper[4824]: E1209 10:17:51.982912 4824 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-7d54cc949d-rzfx4" podUID="d8154a7d-1dbc-4fa3-a862-140f2fd807b2" containerName="heat-engine" Dec 09 10:17:51 crc kubenswrapper[4824]: I1209 10:17:51.991859 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-bpqmt"] Dec 09 10:17:51 crc kubenswrapper[4824]: E1209 10:17:51.992412 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6e9dedb-7873-4008-a9a0-fa90ec7c99f1" containerName="heat-api" Dec 09 10:17:51 crc kubenswrapper[4824]: I1209 10:17:51.992424 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6e9dedb-7873-4008-a9a0-fa90ec7c99f1" containerName="heat-api" Dec 09 10:17:51 crc kubenswrapper[4824]: E1209 10:17:51.992475 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14358ed8-0c10-48f7-a1b1-37e75e7a6400" containerName="heat-cfnapi" Dec 09 10:17:51 crc kubenswrapper[4824]: I1209 10:17:51.992481 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="14358ed8-0c10-48f7-a1b1-37e75e7a6400" containerName="heat-cfnapi" Dec 09 10:17:51 crc kubenswrapper[4824]: I1209 10:17:51.992706 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="14358ed8-0c10-48f7-a1b1-37e75e7a6400" containerName="heat-cfnapi" Dec 09 10:17:51 crc kubenswrapper[4824]: I1209 10:17:51.992731 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6e9dedb-7873-4008-a9a0-fa90ec7c99f1" containerName="heat-api" Dec 09 10:17:51 crc kubenswrapper[4824]: I1209 10:17:51.993614 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-bpqmt" Dec 09 10:17:52 crc kubenswrapper[4824]: I1209 10:17:52.011319 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-bpqmt"] Dec 09 10:17:52 crc kubenswrapper[4824]: I1209 10:17:52.014083 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 09 10:17:52 crc kubenswrapper[4824]: I1209 10:17:52.062575 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e293dba-63b1-4841-b409-0271f0329251-config-data\") pod \"aodh-db-sync-bpqmt\" (UID: \"3e293dba-63b1-4841-b409-0271f0329251\") " pod="openstack/aodh-db-sync-bpqmt" Dec 09 10:17:52 crc kubenswrapper[4824]: I1209 10:17:52.081199 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e293dba-63b1-4841-b409-0271f0329251-combined-ca-bundle\") pod \"aodh-db-sync-bpqmt\" (UID: \"3e293dba-63b1-4841-b409-0271f0329251\") " pod="openstack/aodh-db-sync-bpqmt" Dec 09 10:17:52 crc kubenswrapper[4824]: I1209 10:17:52.082447 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8zqz\" (UniqueName: \"kubernetes.io/projected/3e293dba-63b1-4841-b409-0271f0329251-kube-api-access-s8zqz\") pod \"aodh-db-sync-bpqmt\" (UID: \"3e293dba-63b1-4841-b409-0271f0329251\") " pod="openstack/aodh-db-sync-bpqmt" Dec 09 10:17:52 crc kubenswrapper[4824]: I1209 10:17:52.082537 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e293dba-63b1-4841-b409-0271f0329251-scripts\") pod \"aodh-db-sync-bpqmt\" (UID: \"3e293dba-63b1-4841-b409-0271f0329251\") " pod="openstack/aodh-db-sync-bpqmt" Dec 09 10:17:52 crc kubenswrapper[4824]: I1209 10:17:52.184939 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8zqz\" (UniqueName: \"kubernetes.io/projected/3e293dba-63b1-4841-b409-0271f0329251-kube-api-access-s8zqz\") pod \"aodh-db-sync-bpqmt\" (UID: \"3e293dba-63b1-4841-b409-0271f0329251\") " pod="openstack/aodh-db-sync-bpqmt" Dec 09 10:17:52 crc kubenswrapper[4824]: I1209 10:17:52.184996 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e293dba-63b1-4841-b409-0271f0329251-scripts\") pod \"aodh-db-sync-bpqmt\" (UID: \"3e293dba-63b1-4841-b409-0271f0329251\") " pod="openstack/aodh-db-sync-bpqmt" Dec 09 10:17:52 crc kubenswrapper[4824]: I1209 10:17:52.185230 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e293dba-63b1-4841-b409-0271f0329251-config-data\") pod \"aodh-db-sync-bpqmt\" (UID: \"3e293dba-63b1-4841-b409-0271f0329251\") " pod="openstack/aodh-db-sync-bpqmt" Dec 09 10:17:52 crc kubenswrapper[4824]: I1209 10:17:52.185275 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e293dba-63b1-4841-b409-0271f0329251-combined-ca-bundle\") pod \"aodh-db-sync-bpqmt\" (UID: \"3e293dba-63b1-4841-b409-0271f0329251\") " pod="openstack/aodh-db-sync-bpqmt" Dec 09 10:17:52 crc kubenswrapper[4824]: I1209 10:17:52.196531 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e293dba-63b1-4841-b409-0271f0329251-scripts\") pod \"aodh-db-sync-bpqmt\" (UID: \"3e293dba-63b1-4841-b409-0271f0329251\") " pod="openstack/aodh-db-sync-bpqmt" Dec 09 10:17:52 crc kubenswrapper[4824]: I1209 10:17:52.196577 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e293dba-63b1-4841-b409-0271f0329251-combined-ca-bundle\") pod \"aodh-db-sync-bpqmt\" (UID: \"3e293dba-63b1-4841-b409-0271f0329251\") " pod="openstack/aodh-db-sync-bpqmt" Dec 09 10:17:52 crc kubenswrapper[4824]: I1209 10:17:52.207447 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e293dba-63b1-4841-b409-0271f0329251-config-data\") pod \"aodh-db-sync-bpqmt\" (UID: \"3e293dba-63b1-4841-b409-0271f0329251\") " pod="openstack/aodh-db-sync-bpqmt" Dec 09 10:17:52 crc kubenswrapper[4824]: I1209 10:17:52.207973 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8zqz\" (UniqueName: \"kubernetes.io/projected/3e293dba-63b1-4841-b409-0271f0329251-kube-api-access-s8zqz\") pod \"aodh-db-sync-bpqmt\" (UID: \"3e293dba-63b1-4841-b409-0271f0329251\") " pod="openstack/aodh-db-sync-bpqmt" Dec 09 10:17:52 crc kubenswrapper[4824]: I1209 10:17:52.346855 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-bpqmt" Dec 09 10:17:53 crc kubenswrapper[4824]: I1209 10:17:53.969155 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4eb5fcc0-0113-4709-881b-7d109571a08d" path="/var/lib/kubelet/pods/4eb5fcc0-0113-4709-881b-7d109571a08d/volumes" Dec 09 10:17:54 crc kubenswrapper[4824]: I1209 10:17:54.350816 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-bpqmt"] Dec 09 10:17:55 crc kubenswrapper[4824]: I1209 10:17:55.006984 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-bpqmt" event={"ID":"3e293dba-63b1-4841-b409-0271f0329251","Type":"ContainerStarted","Data":"01316c08d91b7f887f26f5a485475b26c7555d36ae247aaf57ba671122d713be"} Dec 09 10:17:55 crc kubenswrapper[4824]: I1209 10:17:55.009192 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j" event={"ID":"6e1db872-7d04-4c89-b725-9e3d84fed0fb","Type":"ContainerStarted","Data":"edc4072b05f90590594b9131d4656f94125f260d1dcc9d3e52f7cf39c0ec27cf"} Dec 09 10:17:55 crc kubenswrapper[4824]: I1209 10:17:55.030160 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j" podStartSLOduration=3.444347282 podStartE2EDuration="18.030143631s" podCreationTimestamp="2025-12-09 10:17:37 +0000 UTC" firstStartedPulling="2025-12-09 10:17:39.092679681 +0000 UTC m=+1815.427184348" lastFinishedPulling="2025-12-09 10:17:53.67847603 +0000 UTC m=+1830.012980697" observedRunningTime="2025-12-09 10:17:55.0253462 +0000 UTC m=+1831.359850867" watchObservedRunningTime="2025-12-09 10:17:55.030143631 +0000 UTC m=+1831.364648298" Dec 09 10:17:57 crc kubenswrapper[4824]: I1209 10:17:57.872618 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 09 10:17:58 crc kubenswrapper[4824]: I1209 10:17:58.582013 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 09 10:17:59 crc kubenswrapper[4824]: I1209 10:17:59.910865 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:17:59 crc kubenswrapper[4824]: E1209 10:17:59.911634 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:18:00 crc kubenswrapper[4824]: I1209 10:18:00.081695 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-bpqmt" event={"ID":"3e293dba-63b1-4841-b409-0271f0329251","Type":"ContainerStarted","Data":"a9c3600fe3735139a42bd8d826c49d7371117ce269bbb7d36569e4497ca5a539"} Dec 09 10:18:00 crc kubenswrapper[4824]: I1209 10:18:00.114460 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-bpqmt" podStartSLOduration=3.972421071 podStartE2EDuration="9.114432772s" podCreationTimestamp="2025-12-09 10:17:51 +0000 UTC" firstStartedPulling="2025-12-09 10:17:54.352688069 +0000 UTC m=+1830.687192736" lastFinishedPulling="2025-12-09 10:17:59.49469977 +0000 UTC m=+1835.829204437" observedRunningTime="2025-12-09 10:18:00.100013278 +0000 UTC m=+1836.434517955" watchObservedRunningTime="2025-12-09 10:18:00.114432772 +0000 UTC m=+1836.448937439" Dec 09 10:18:01 crc kubenswrapper[4824]: E1209 10:18:01.950169 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6a711400741bf3c06a5572c6a46ac91b8ec31fb5fdfda32e199126f0b080b249" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 09 10:18:01 crc kubenswrapper[4824]: E1209 10:18:01.951945 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6a711400741bf3c06a5572c6a46ac91b8ec31fb5fdfda32e199126f0b080b249" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 09 10:18:01 crc kubenswrapper[4824]: E1209 10:18:01.954757 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6a711400741bf3c06a5572c6a46ac91b8ec31fb5fdfda32e199126f0b080b249" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 09 10:18:01 crc kubenswrapper[4824]: E1209 10:18:01.954827 4824 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-7d54cc949d-rzfx4" podUID="d8154a7d-1dbc-4fa3-a862-140f2fd807b2" containerName="heat-engine" Dec 09 10:18:03 crc kubenswrapper[4824]: I1209 10:18:03.120984 4824 generic.go:334] "Generic (PLEG): container finished" podID="3e293dba-63b1-4841-b409-0271f0329251" containerID="a9c3600fe3735139a42bd8d826c49d7371117ce269bbb7d36569e4497ca5a539" exitCode=0 Dec 09 10:18:03 crc kubenswrapper[4824]: I1209 10:18:03.121319 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-bpqmt" event={"ID":"3e293dba-63b1-4841-b409-0271f0329251","Type":"ContainerDied","Data":"a9c3600fe3735139a42bd8d826c49d7371117ce269bbb7d36569e4497ca5a539"} Dec 09 10:18:04 crc kubenswrapper[4824]: I1209 10:18:04.626205 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-bpqmt" Dec 09 10:18:04 crc kubenswrapper[4824]: I1209 10:18:04.731902 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8zqz\" (UniqueName: \"kubernetes.io/projected/3e293dba-63b1-4841-b409-0271f0329251-kube-api-access-s8zqz\") pod \"3e293dba-63b1-4841-b409-0271f0329251\" (UID: \"3e293dba-63b1-4841-b409-0271f0329251\") " Dec 09 10:18:04 crc kubenswrapper[4824]: I1209 10:18:04.732081 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e293dba-63b1-4841-b409-0271f0329251-config-data\") pod \"3e293dba-63b1-4841-b409-0271f0329251\" (UID: \"3e293dba-63b1-4841-b409-0271f0329251\") " Dec 09 10:18:04 crc kubenswrapper[4824]: I1209 10:18:04.732169 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e293dba-63b1-4841-b409-0271f0329251-combined-ca-bundle\") pod \"3e293dba-63b1-4841-b409-0271f0329251\" (UID: \"3e293dba-63b1-4841-b409-0271f0329251\") " Dec 09 10:18:04 crc kubenswrapper[4824]: I1209 10:18:04.732217 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e293dba-63b1-4841-b409-0271f0329251-scripts\") pod \"3e293dba-63b1-4841-b409-0271f0329251\" (UID: \"3e293dba-63b1-4841-b409-0271f0329251\") " Dec 09 10:18:04 crc kubenswrapper[4824]: I1209 10:18:04.738964 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e293dba-63b1-4841-b409-0271f0329251-kube-api-access-s8zqz" (OuterVolumeSpecName: "kube-api-access-s8zqz") pod "3e293dba-63b1-4841-b409-0271f0329251" (UID: "3e293dba-63b1-4841-b409-0271f0329251"). InnerVolumeSpecName "kube-api-access-s8zqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:18:04 crc kubenswrapper[4824]: I1209 10:18:04.740381 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e293dba-63b1-4841-b409-0271f0329251-scripts" (OuterVolumeSpecName: "scripts") pod "3e293dba-63b1-4841-b409-0271f0329251" (UID: "3e293dba-63b1-4841-b409-0271f0329251"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:18:04 crc kubenswrapper[4824]: I1209 10:18:04.773472 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e293dba-63b1-4841-b409-0271f0329251-config-data" (OuterVolumeSpecName: "config-data") pod "3e293dba-63b1-4841-b409-0271f0329251" (UID: "3e293dba-63b1-4841-b409-0271f0329251"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:18:04 crc kubenswrapper[4824]: I1209 10:18:04.775606 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e293dba-63b1-4841-b409-0271f0329251-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3e293dba-63b1-4841-b409-0271f0329251" (UID: "3e293dba-63b1-4841-b409-0271f0329251"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:18:04 crc kubenswrapper[4824]: I1209 10:18:04.836142 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8zqz\" (UniqueName: \"kubernetes.io/projected/3e293dba-63b1-4841-b409-0271f0329251-kube-api-access-s8zqz\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:04 crc kubenswrapper[4824]: I1209 10:18:04.836179 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e293dba-63b1-4841-b409-0271f0329251-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:04 crc kubenswrapper[4824]: I1209 10:18:04.836193 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e293dba-63b1-4841-b409-0271f0329251-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:04 crc kubenswrapper[4824]: I1209 10:18:04.836205 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e293dba-63b1-4841-b409-0271f0329251-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:05 crc kubenswrapper[4824]: I1209 10:18:05.165122 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-bpqmt" event={"ID":"3e293dba-63b1-4841-b409-0271f0329251","Type":"ContainerDied","Data":"01316c08d91b7f887f26f5a485475b26c7555d36ae247aaf57ba671122d713be"} Dec 09 10:18:05 crc kubenswrapper[4824]: I1209 10:18:05.165390 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01316c08d91b7f887f26f5a485475b26c7555d36ae247aaf57ba671122d713be" Dec 09 10:18:05 crc kubenswrapper[4824]: I1209 10:18:05.165471 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-bpqmt" Dec 09 10:18:06 crc kubenswrapper[4824]: I1209 10:18:06.181966 4824 generic.go:334] "Generic (PLEG): container finished" podID="6e1db872-7d04-4c89-b725-9e3d84fed0fb" containerID="edc4072b05f90590594b9131d4656f94125f260d1dcc9d3e52f7cf39c0ec27cf" exitCode=0 Dec 09 10:18:06 crc kubenswrapper[4824]: I1209 10:18:06.182072 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j" event={"ID":"6e1db872-7d04-4c89-b725-9e3d84fed0fb","Type":"ContainerDied","Data":"edc4072b05f90590594b9131d4656f94125f260d1dcc9d3e52f7cf39c0ec27cf"} Dec 09 10:18:06 crc kubenswrapper[4824]: I1209 10:18:06.187592 4824 generic.go:334] "Generic (PLEG): container finished" podID="d8154a7d-1dbc-4fa3-a862-140f2fd807b2" containerID="6a711400741bf3c06a5572c6a46ac91b8ec31fb5fdfda32e199126f0b080b249" exitCode=0 Dec 09 10:18:06 crc kubenswrapper[4824]: I1209 10:18:06.187650 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7d54cc949d-rzfx4" event={"ID":"d8154a7d-1dbc-4fa3-a862-140f2fd807b2","Type":"ContainerDied","Data":"6a711400741bf3c06a5572c6a46ac91b8ec31fb5fdfda32e199126f0b080b249"} Dec 09 10:18:06 crc kubenswrapper[4824]: I1209 10:18:06.441494 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7d54cc949d-rzfx4" Dec 09 10:18:06 crc kubenswrapper[4824]: I1209 10:18:06.581344 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8154a7d-1dbc-4fa3-a862-140f2fd807b2-config-data-custom\") pod \"d8154a7d-1dbc-4fa3-a862-140f2fd807b2\" (UID: \"d8154a7d-1dbc-4fa3-a862-140f2fd807b2\") " Dec 09 10:18:06 crc kubenswrapper[4824]: I1209 10:18:06.581433 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8154a7d-1dbc-4fa3-a862-140f2fd807b2-combined-ca-bundle\") pod \"d8154a7d-1dbc-4fa3-a862-140f2fd807b2\" (UID: \"d8154a7d-1dbc-4fa3-a862-140f2fd807b2\") " Dec 09 10:18:06 crc kubenswrapper[4824]: I1209 10:18:06.581462 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rw8r\" (UniqueName: \"kubernetes.io/projected/d8154a7d-1dbc-4fa3-a862-140f2fd807b2-kube-api-access-7rw8r\") pod \"d8154a7d-1dbc-4fa3-a862-140f2fd807b2\" (UID: \"d8154a7d-1dbc-4fa3-a862-140f2fd807b2\") " Dec 09 10:18:06 crc kubenswrapper[4824]: I1209 10:18:06.581596 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8154a7d-1dbc-4fa3-a862-140f2fd807b2-config-data\") pod \"d8154a7d-1dbc-4fa3-a862-140f2fd807b2\" (UID: \"d8154a7d-1dbc-4fa3-a862-140f2fd807b2\") " Dec 09 10:18:06 crc kubenswrapper[4824]: I1209 10:18:06.587565 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8154a7d-1dbc-4fa3-a862-140f2fd807b2-kube-api-access-7rw8r" (OuterVolumeSpecName: "kube-api-access-7rw8r") pod "d8154a7d-1dbc-4fa3-a862-140f2fd807b2" (UID: "d8154a7d-1dbc-4fa3-a862-140f2fd807b2"). InnerVolumeSpecName "kube-api-access-7rw8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:18:06 crc kubenswrapper[4824]: I1209 10:18:06.587663 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8154a7d-1dbc-4fa3-a862-140f2fd807b2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d8154a7d-1dbc-4fa3-a862-140f2fd807b2" (UID: "d8154a7d-1dbc-4fa3-a862-140f2fd807b2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:18:06 crc kubenswrapper[4824]: I1209 10:18:06.623615 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8154a7d-1dbc-4fa3-a862-140f2fd807b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d8154a7d-1dbc-4fa3-a862-140f2fd807b2" (UID: "d8154a7d-1dbc-4fa3-a862-140f2fd807b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:18:06 crc kubenswrapper[4824]: I1209 10:18:06.654110 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8154a7d-1dbc-4fa3-a862-140f2fd807b2-config-data" (OuterVolumeSpecName: "config-data") pod "d8154a7d-1dbc-4fa3-a862-140f2fd807b2" (UID: "d8154a7d-1dbc-4fa3-a862-140f2fd807b2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:18:06 crc kubenswrapper[4824]: I1209 10:18:06.685073 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8154a7d-1dbc-4fa3-a862-140f2fd807b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:06 crc kubenswrapper[4824]: I1209 10:18:06.685109 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rw8r\" (UniqueName: \"kubernetes.io/projected/d8154a7d-1dbc-4fa3-a862-140f2fd807b2-kube-api-access-7rw8r\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:06 crc kubenswrapper[4824]: I1209 10:18:06.685126 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8154a7d-1dbc-4fa3-a862-140f2fd807b2-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:06 crc kubenswrapper[4824]: I1209 10:18:06.685137 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8154a7d-1dbc-4fa3-a862-140f2fd807b2-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:06 crc kubenswrapper[4824]: I1209 10:18:06.957183 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 09 10:18:06 crc kubenswrapper[4824]: I1209 10:18:06.957743 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" containerName="aodh-api" containerID="cri-o://996c1915e4732b450d51c25f2a00141f01516d9c5baedbaa1da35c6a1ffe3be2" gracePeriod=30 Dec 09 10:18:06 crc kubenswrapper[4824]: I1209 10:18:06.957843 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" containerName="aodh-listener" containerID="cri-o://fb69b48a84b960f40b8160e2b122bf60e8d692316bc07251392c08bfa27737b6" gracePeriod=30 Dec 09 10:18:06 crc kubenswrapper[4824]: I1209 10:18:06.957900 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" containerName="aodh-evaluator" containerID="cri-o://990c7f6cffbc366e0e2c8772c94ef0716443d4bfc59f04eb74f5fd264c37e9a2" gracePeriod=30 Dec 09 10:18:06 crc kubenswrapper[4824]: I1209 10:18:06.957934 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" containerName="aodh-notifier" containerID="cri-o://92c378f182312eddc1cf9b26ea46a7f2ca06588876fda5a793d6e23092f4d12a" gracePeriod=30 Dec 09 10:18:07 crc kubenswrapper[4824]: I1209 10:18:07.213480 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7d54cc949d-rzfx4" event={"ID":"d8154a7d-1dbc-4fa3-a862-140f2fd807b2","Type":"ContainerDied","Data":"deaf4df5a37b1d42169f79c4f2bb9cacd88e9626a50040e4e637af0427f8ad8c"} Dec 09 10:18:07 crc kubenswrapper[4824]: I1209 10:18:07.213552 4824 scope.go:117] "RemoveContainer" containerID="6a711400741bf3c06a5572c6a46ac91b8ec31fb5fdfda32e199126f0b080b249" Dec 09 10:18:07 crc kubenswrapper[4824]: I1209 10:18:07.213728 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7d54cc949d-rzfx4" Dec 09 10:18:07 crc kubenswrapper[4824]: I1209 10:18:07.228293 4824 generic.go:334] "Generic (PLEG): container finished" podID="d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" containerID="996c1915e4732b450d51c25f2a00141f01516d9c5baedbaa1da35c6a1ffe3be2" exitCode=0 Dec 09 10:18:07 crc kubenswrapper[4824]: I1209 10:18:07.228682 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912","Type":"ContainerDied","Data":"996c1915e4732b450d51c25f2a00141f01516d9c5baedbaa1da35c6a1ffe3be2"} Dec 09 10:18:07 crc kubenswrapper[4824]: I1209 10:18:07.278701 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-7d54cc949d-rzfx4"] Dec 09 10:18:07 crc kubenswrapper[4824]: I1209 10:18:07.292642 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-7d54cc949d-rzfx4"] Dec 09 10:18:07 crc kubenswrapper[4824]: I1209 10:18:07.905711 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j" Dec 09 10:18:07 crc kubenswrapper[4824]: I1209 10:18:07.924638 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8154a7d-1dbc-4fa3-a862-140f2fd807b2" path="/var/lib/kubelet/pods/d8154a7d-1dbc-4fa3-a862-140f2fd807b2/volumes" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.030033 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e1db872-7d04-4c89-b725-9e3d84fed0fb-inventory\") pod \"6e1db872-7d04-4c89-b725-9e3d84fed0fb\" (UID: \"6e1db872-7d04-4c89-b725-9e3d84fed0fb\") " Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.030118 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e1db872-7d04-4c89-b725-9e3d84fed0fb-ssh-key\") pod \"6e1db872-7d04-4c89-b725-9e3d84fed0fb\" (UID: \"6e1db872-7d04-4c89-b725-9e3d84fed0fb\") " Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.030363 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e1db872-7d04-4c89-b725-9e3d84fed0fb-repo-setup-combined-ca-bundle\") pod \"6e1db872-7d04-4c89-b725-9e3d84fed0fb\" (UID: \"6e1db872-7d04-4c89-b725-9e3d84fed0fb\") " Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.030431 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tp4fh\" (UniqueName: \"kubernetes.io/projected/6e1db872-7d04-4c89-b725-9e3d84fed0fb-kube-api-access-tp4fh\") pod \"6e1db872-7d04-4c89-b725-9e3d84fed0fb\" (UID: \"6e1db872-7d04-4c89-b725-9e3d84fed0fb\") " Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.036298 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e1db872-7d04-4c89-b725-9e3d84fed0fb-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "6e1db872-7d04-4c89-b725-9e3d84fed0fb" (UID: "6e1db872-7d04-4c89-b725-9e3d84fed0fb"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.037803 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e1db872-7d04-4c89-b725-9e3d84fed0fb-kube-api-access-tp4fh" (OuterVolumeSpecName: "kube-api-access-tp4fh") pod "6e1db872-7d04-4c89-b725-9e3d84fed0fb" (UID: "6e1db872-7d04-4c89-b725-9e3d84fed0fb"). InnerVolumeSpecName "kube-api-access-tp4fh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.071941 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e1db872-7d04-4c89-b725-9e3d84fed0fb-inventory" (OuterVolumeSpecName: "inventory") pod "6e1db872-7d04-4c89-b725-9e3d84fed0fb" (UID: "6e1db872-7d04-4c89-b725-9e3d84fed0fb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.073670 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e1db872-7d04-4c89-b725-9e3d84fed0fb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6e1db872-7d04-4c89-b725-9e3d84fed0fb" (UID: "6e1db872-7d04-4c89-b725-9e3d84fed0fb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.133636 4824 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e1db872-7d04-4c89-b725-9e3d84fed0fb-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.133680 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tp4fh\" (UniqueName: \"kubernetes.io/projected/6e1db872-7d04-4c89-b725-9e3d84fed0fb-kube-api-access-tp4fh\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.133702 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e1db872-7d04-4c89-b725-9e3d84fed0fb-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.133715 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e1db872-7d04-4c89-b725-9e3d84fed0fb-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.243984 4824 generic.go:334] "Generic (PLEG): container finished" podID="d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" containerID="92c378f182312eddc1cf9b26ea46a7f2ca06588876fda5a793d6e23092f4d12a" exitCode=0 Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.245020 4824 generic.go:334] "Generic (PLEG): container finished" podID="d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" containerID="990c7f6cffbc366e0e2c8772c94ef0716443d4bfc59f04eb74f5fd264c37e9a2" exitCode=0 Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.244980 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912","Type":"ContainerDied","Data":"92c378f182312eddc1cf9b26ea46a7f2ca06588876fda5a793d6e23092f4d12a"} Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.245210 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912","Type":"ContainerDied","Data":"990c7f6cffbc366e0e2c8772c94ef0716443d4bfc59f04eb74f5fd264c37e9a2"} Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.248537 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j" event={"ID":"6e1db872-7d04-4c89-b725-9e3d84fed0fb","Type":"ContainerDied","Data":"32431b06114800a867d9bd7c8974545de76e793a5c57384fffe2f03d5a602396"} Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.248837 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32431b06114800a867d9bd7c8974545de76e793a5c57384fffe2f03d5a602396" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.248805 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.301368 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-vbfns"] Dec 09 10:18:08 crc kubenswrapper[4824]: E1209 10:18:08.302067 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8154a7d-1dbc-4fa3-a862-140f2fd807b2" containerName="heat-engine" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.302089 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8154a7d-1dbc-4fa3-a862-140f2fd807b2" containerName="heat-engine" Dec 09 10:18:08 crc kubenswrapper[4824]: E1209 10:18:08.302136 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e293dba-63b1-4841-b409-0271f0329251" containerName="aodh-db-sync" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.302147 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e293dba-63b1-4841-b409-0271f0329251" containerName="aodh-db-sync" Dec 09 10:18:08 crc kubenswrapper[4824]: E1209 10:18:08.302166 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e1db872-7d04-4c89-b725-9e3d84fed0fb" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.302176 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e1db872-7d04-4c89-b725-9e3d84fed0fb" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.302496 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e1db872-7d04-4c89-b725-9e3d84fed0fb" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.302534 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8154a7d-1dbc-4fa3-a862-140f2fd807b2" containerName="heat-engine" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.302552 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e293dba-63b1-4841-b409-0271f0329251" containerName="aodh-db-sync" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.303687 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vbfns" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.307076 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.307461 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.307661 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5nzld" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.307847 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.319233 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-vbfns"] Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.440266 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eeb8228d-77ed-45ad-b2dc-085a2ba8814e-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vbfns\" (UID: \"eeb8228d-77ed-45ad-b2dc-085a2ba8814e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vbfns" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.440467 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eeb8228d-77ed-45ad-b2dc-085a2ba8814e-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vbfns\" (UID: \"eeb8228d-77ed-45ad-b2dc-085a2ba8814e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vbfns" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.440731 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kghr\" (UniqueName: \"kubernetes.io/projected/eeb8228d-77ed-45ad-b2dc-085a2ba8814e-kube-api-access-9kghr\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vbfns\" (UID: \"eeb8228d-77ed-45ad-b2dc-085a2ba8814e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vbfns" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.543481 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kghr\" (UniqueName: \"kubernetes.io/projected/eeb8228d-77ed-45ad-b2dc-085a2ba8814e-kube-api-access-9kghr\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vbfns\" (UID: \"eeb8228d-77ed-45ad-b2dc-085a2ba8814e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vbfns" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.543959 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eeb8228d-77ed-45ad-b2dc-085a2ba8814e-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vbfns\" (UID: \"eeb8228d-77ed-45ad-b2dc-085a2ba8814e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vbfns" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.544123 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eeb8228d-77ed-45ad-b2dc-085a2ba8814e-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vbfns\" (UID: \"eeb8228d-77ed-45ad-b2dc-085a2ba8814e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vbfns" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.549914 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eeb8228d-77ed-45ad-b2dc-085a2ba8814e-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vbfns\" (UID: \"eeb8228d-77ed-45ad-b2dc-085a2ba8814e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vbfns" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.558059 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eeb8228d-77ed-45ad-b2dc-085a2ba8814e-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vbfns\" (UID: \"eeb8228d-77ed-45ad-b2dc-085a2ba8814e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vbfns" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.560643 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kghr\" (UniqueName: \"kubernetes.io/projected/eeb8228d-77ed-45ad-b2dc-085a2ba8814e-kube-api-access-9kghr\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vbfns\" (UID: \"eeb8228d-77ed-45ad-b2dc-085a2ba8814e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vbfns" Dec 09 10:18:08 crc kubenswrapper[4824]: E1209 10:18:08.569547 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e1db872_7d04_4c89_b725_9e3d84fed0fb.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e1db872_7d04_4c89_b725_9e3d84fed0fb.slice/crio-32431b06114800a867d9bd7c8974545de76e793a5c57384fffe2f03d5a602396\": RecentStats: unable to find data in memory cache]" Dec 09 10:18:08 crc kubenswrapper[4824]: I1209 10:18:08.629584 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vbfns" Dec 09 10:18:09 crc kubenswrapper[4824]: I1209 10:18:09.248810 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-vbfns"] Dec 09 10:18:09 crc kubenswrapper[4824]: I1209 10:18:09.262198 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vbfns" event={"ID":"eeb8228d-77ed-45ad-b2dc-085a2ba8814e","Type":"ContainerStarted","Data":"9a35c5c35cbb4abc8e5997e57c9a664d25ee72562df278069b12eecc5aa5d25f"} Dec 09 10:18:10 crc kubenswrapper[4824]: I1209 10:18:10.911554 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:18:10 crc kubenswrapper[4824]: E1209 10:18:10.912664 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.217498 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.335881 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-combined-ca-bundle\") pod \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\" (UID: \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\") " Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.336199 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-internal-tls-certs\") pod \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\" (UID: \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\") " Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.336227 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-config-data\") pod \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\" (UID: \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\") " Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.336258 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbxp8\" (UniqueName: \"kubernetes.io/projected/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-kube-api-access-mbxp8\") pod \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\" (UID: \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\") " Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.336294 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-scripts\") pod \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\" (UID: \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\") " Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.336364 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-public-tls-certs\") pod \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\" (UID: \"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912\") " Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.361281 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-scripts" (OuterVolumeSpecName: "scripts") pod "d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" (UID: "d9d7c93f-e88e-4b55-91b8-f9dc9f42b912"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.370967 4824 generic.go:334] "Generic (PLEG): container finished" podID="d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" containerID="fb69b48a84b960f40b8160e2b122bf60e8d692316bc07251392c08bfa27737b6" exitCode=0 Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.371011 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912","Type":"ContainerDied","Data":"fb69b48a84b960f40b8160e2b122bf60e8d692316bc07251392c08bfa27737b6"} Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.371041 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d9d7c93f-e88e-4b55-91b8-f9dc9f42b912","Type":"ContainerDied","Data":"8d712badb43050a0c7c9b75963f99c41b35c44edc9ac1c5bd070f05ad7b5dc6c"} Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.371063 4824 scope.go:117] "RemoveContainer" containerID="fb69b48a84b960f40b8160e2b122bf60e8d692316bc07251392c08bfa27737b6" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.371301 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.372086 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-kube-api-access-mbxp8" (OuterVolumeSpecName: "kube-api-access-mbxp8") pod "d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" (UID: "d9d7c93f-e88e-4b55-91b8-f9dc9f42b912"). InnerVolumeSpecName "kube-api-access-mbxp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.442176 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbxp8\" (UniqueName: \"kubernetes.io/projected/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-kube-api-access-mbxp8\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.442205 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.542970 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" (UID: "d9d7c93f-e88e-4b55-91b8-f9dc9f42b912"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.552844 4824 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.570886 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" (UID: "d9d7c93f-e88e-4b55-91b8-f9dc9f42b912"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.588111 4824 scope.go:117] "RemoveContainer" containerID="92c378f182312eddc1cf9b26ea46a7f2ca06588876fda5a793d6e23092f4d12a" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.624250 4824 scope.go:117] "RemoveContainer" containerID="990c7f6cffbc366e0e2c8772c94ef0716443d4bfc59f04eb74f5fd264c37e9a2" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.634015 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" (UID: "d9d7c93f-e88e-4b55-91b8-f9dc9f42b912"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.651873 4824 scope.go:117] "RemoveContainer" containerID="996c1915e4732b450d51c25f2a00141f01516d9c5baedbaa1da35c6a1ffe3be2" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.654862 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.654885 4824 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.677964 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-config-data" (OuterVolumeSpecName: "config-data") pod "d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" (UID: "d9d7c93f-e88e-4b55-91b8-f9dc9f42b912"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.682176 4824 scope.go:117] "RemoveContainer" containerID="fb69b48a84b960f40b8160e2b122bf60e8d692316bc07251392c08bfa27737b6" Dec 09 10:18:11 crc kubenswrapper[4824]: E1209 10:18:11.683050 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb69b48a84b960f40b8160e2b122bf60e8d692316bc07251392c08bfa27737b6\": container with ID starting with fb69b48a84b960f40b8160e2b122bf60e8d692316bc07251392c08bfa27737b6 not found: ID does not exist" containerID="fb69b48a84b960f40b8160e2b122bf60e8d692316bc07251392c08bfa27737b6" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.683098 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb69b48a84b960f40b8160e2b122bf60e8d692316bc07251392c08bfa27737b6"} err="failed to get container status \"fb69b48a84b960f40b8160e2b122bf60e8d692316bc07251392c08bfa27737b6\": rpc error: code = NotFound desc = could not find container \"fb69b48a84b960f40b8160e2b122bf60e8d692316bc07251392c08bfa27737b6\": container with ID starting with fb69b48a84b960f40b8160e2b122bf60e8d692316bc07251392c08bfa27737b6 not found: ID does not exist" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.683124 4824 scope.go:117] "RemoveContainer" containerID="92c378f182312eddc1cf9b26ea46a7f2ca06588876fda5a793d6e23092f4d12a" Dec 09 10:18:11 crc kubenswrapper[4824]: E1209 10:18:11.683431 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92c378f182312eddc1cf9b26ea46a7f2ca06588876fda5a793d6e23092f4d12a\": container with ID starting with 92c378f182312eddc1cf9b26ea46a7f2ca06588876fda5a793d6e23092f4d12a not found: ID does not exist" containerID="92c378f182312eddc1cf9b26ea46a7f2ca06588876fda5a793d6e23092f4d12a" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.683460 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92c378f182312eddc1cf9b26ea46a7f2ca06588876fda5a793d6e23092f4d12a"} err="failed to get container status \"92c378f182312eddc1cf9b26ea46a7f2ca06588876fda5a793d6e23092f4d12a\": rpc error: code = NotFound desc = could not find container \"92c378f182312eddc1cf9b26ea46a7f2ca06588876fda5a793d6e23092f4d12a\": container with ID starting with 92c378f182312eddc1cf9b26ea46a7f2ca06588876fda5a793d6e23092f4d12a not found: ID does not exist" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.683480 4824 scope.go:117] "RemoveContainer" containerID="990c7f6cffbc366e0e2c8772c94ef0716443d4bfc59f04eb74f5fd264c37e9a2" Dec 09 10:18:11 crc kubenswrapper[4824]: E1209 10:18:11.683702 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"990c7f6cffbc366e0e2c8772c94ef0716443d4bfc59f04eb74f5fd264c37e9a2\": container with ID starting with 990c7f6cffbc366e0e2c8772c94ef0716443d4bfc59f04eb74f5fd264c37e9a2 not found: ID does not exist" containerID="990c7f6cffbc366e0e2c8772c94ef0716443d4bfc59f04eb74f5fd264c37e9a2" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.683721 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"990c7f6cffbc366e0e2c8772c94ef0716443d4bfc59f04eb74f5fd264c37e9a2"} err="failed to get container status \"990c7f6cffbc366e0e2c8772c94ef0716443d4bfc59f04eb74f5fd264c37e9a2\": rpc error: code = NotFound desc = could not find container \"990c7f6cffbc366e0e2c8772c94ef0716443d4bfc59f04eb74f5fd264c37e9a2\": container with ID starting with 990c7f6cffbc366e0e2c8772c94ef0716443d4bfc59f04eb74f5fd264c37e9a2 not found: ID does not exist" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.683735 4824 scope.go:117] "RemoveContainer" containerID="996c1915e4732b450d51c25f2a00141f01516d9c5baedbaa1da35c6a1ffe3be2" Dec 09 10:18:11 crc kubenswrapper[4824]: E1209 10:18:11.684304 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"996c1915e4732b450d51c25f2a00141f01516d9c5baedbaa1da35c6a1ffe3be2\": container with ID starting with 996c1915e4732b450d51c25f2a00141f01516d9c5baedbaa1da35c6a1ffe3be2 not found: ID does not exist" containerID="996c1915e4732b450d51c25f2a00141f01516d9c5baedbaa1da35c6a1ffe3be2" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.684332 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"996c1915e4732b450d51c25f2a00141f01516d9c5baedbaa1da35c6a1ffe3be2"} err="failed to get container status \"996c1915e4732b450d51c25f2a00141f01516d9c5baedbaa1da35c6a1ffe3be2\": rpc error: code = NotFound desc = could not find container \"996c1915e4732b450d51c25f2a00141f01516d9c5baedbaa1da35c6a1ffe3be2\": container with ID starting with 996c1915e4732b450d51c25f2a00141f01516d9c5baedbaa1da35c6a1ffe3be2 not found: ID does not exist" Dec 09 10:18:11 crc kubenswrapper[4824]: I1209 10:18:11.757926 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.007539 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.023493 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.053593 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 09 10:18:12 crc kubenswrapper[4824]: E1209 10:18:12.064492 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" containerName="aodh-api" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.064532 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" containerName="aodh-api" Dec 09 10:18:12 crc kubenswrapper[4824]: E1209 10:18:12.064636 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" containerName="aodh-evaluator" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.064646 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" containerName="aodh-evaluator" Dec 09 10:18:12 crc kubenswrapper[4824]: E1209 10:18:12.064679 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" containerName="aodh-notifier" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.064688 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" containerName="aodh-notifier" Dec 09 10:18:12 crc kubenswrapper[4824]: E1209 10:18:12.064732 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" containerName="aodh-listener" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.064741 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" containerName="aodh-listener" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.065643 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" containerName="aodh-listener" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.065663 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" containerName="aodh-evaluator" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.065691 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" containerName="aodh-api" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.065709 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" containerName="aodh-notifier" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.102907 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.104187 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.107121 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.107305 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.107760 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-mgzrr" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.108051 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.108145 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.186059 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/af7caa74-1882-4e75-9c52-942df2e40efe-internal-tls-certs\") pod \"aodh-0\" (UID: \"af7caa74-1882-4e75-9c52-942df2e40efe\") " pod="openstack/aodh-0" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.186168 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8v49\" (UniqueName: \"kubernetes.io/projected/af7caa74-1882-4e75-9c52-942df2e40efe-kube-api-access-s8v49\") pod \"aodh-0\" (UID: \"af7caa74-1882-4e75-9c52-942df2e40efe\") " pod="openstack/aodh-0" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.186232 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af7caa74-1882-4e75-9c52-942df2e40efe-scripts\") pod \"aodh-0\" (UID: \"af7caa74-1882-4e75-9c52-942df2e40efe\") " pod="openstack/aodh-0" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.186387 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/af7caa74-1882-4e75-9c52-942df2e40efe-public-tls-certs\") pod \"aodh-0\" (UID: \"af7caa74-1882-4e75-9c52-942df2e40efe\") " pod="openstack/aodh-0" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.186709 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af7caa74-1882-4e75-9c52-942df2e40efe-combined-ca-bundle\") pod \"aodh-0\" (UID: \"af7caa74-1882-4e75-9c52-942df2e40efe\") " pod="openstack/aodh-0" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.186936 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af7caa74-1882-4e75-9c52-942df2e40efe-config-data\") pod \"aodh-0\" (UID: \"af7caa74-1882-4e75-9c52-942df2e40efe\") " pod="openstack/aodh-0" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.289245 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/af7caa74-1882-4e75-9c52-942df2e40efe-internal-tls-certs\") pod \"aodh-0\" (UID: \"af7caa74-1882-4e75-9c52-942df2e40efe\") " pod="openstack/aodh-0" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.289315 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8v49\" (UniqueName: \"kubernetes.io/projected/af7caa74-1882-4e75-9c52-942df2e40efe-kube-api-access-s8v49\") pod \"aodh-0\" (UID: \"af7caa74-1882-4e75-9c52-942df2e40efe\") " pod="openstack/aodh-0" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.289353 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af7caa74-1882-4e75-9c52-942df2e40efe-scripts\") pod \"aodh-0\" (UID: \"af7caa74-1882-4e75-9c52-942df2e40efe\") " pod="openstack/aodh-0" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.289407 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/af7caa74-1882-4e75-9c52-942df2e40efe-public-tls-certs\") pod \"aodh-0\" (UID: \"af7caa74-1882-4e75-9c52-942df2e40efe\") " pod="openstack/aodh-0" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.289501 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af7caa74-1882-4e75-9c52-942df2e40efe-combined-ca-bundle\") pod \"aodh-0\" (UID: \"af7caa74-1882-4e75-9c52-942df2e40efe\") " pod="openstack/aodh-0" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.289597 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af7caa74-1882-4e75-9c52-942df2e40efe-config-data\") pod \"aodh-0\" (UID: \"af7caa74-1882-4e75-9c52-942df2e40efe\") " pod="openstack/aodh-0" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.294863 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/af7caa74-1882-4e75-9c52-942df2e40efe-internal-tls-certs\") pod \"aodh-0\" (UID: \"af7caa74-1882-4e75-9c52-942df2e40efe\") " pod="openstack/aodh-0" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.296538 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af7caa74-1882-4e75-9c52-942df2e40efe-scripts\") pod \"aodh-0\" (UID: \"af7caa74-1882-4e75-9c52-942df2e40efe\") " pod="openstack/aodh-0" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.297417 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/af7caa74-1882-4e75-9c52-942df2e40efe-public-tls-certs\") pod \"aodh-0\" (UID: \"af7caa74-1882-4e75-9c52-942df2e40efe\") " pod="openstack/aodh-0" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.297753 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af7caa74-1882-4e75-9c52-942df2e40efe-config-data\") pod \"aodh-0\" (UID: \"af7caa74-1882-4e75-9c52-942df2e40efe\") " pod="openstack/aodh-0" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.302058 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af7caa74-1882-4e75-9c52-942df2e40efe-combined-ca-bundle\") pod \"aodh-0\" (UID: \"af7caa74-1882-4e75-9c52-942df2e40efe\") " pod="openstack/aodh-0" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.306729 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8v49\" (UniqueName: \"kubernetes.io/projected/af7caa74-1882-4e75-9c52-942df2e40efe-kube-api-access-s8v49\") pod \"aodh-0\" (UID: \"af7caa74-1882-4e75-9c52-942df2e40efe\") " pod="openstack/aodh-0" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.437620 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 09 10:18:12 crc kubenswrapper[4824]: I1209 10:18:12.949170 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 09 10:18:13 crc kubenswrapper[4824]: I1209 10:18:13.400604 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"af7caa74-1882-4e75-9c52-942df2e40efe","Type":"ContainerStarted","Data":"b16ee9c7904f91b3c60ea4dae5f5d89c440041efd85eb31f1b051de7a0170d82"} Dec 09 10:18:13 crc kubenswrapper[4824]: I1209 10:18:13.546958 4824 scope.go:117] "RemoveContainer" containerID="40ff06ab3f0f7bbf01219bb470c28376ccf91ab46f772e08dc41ca02ca67a752" Dec 09 10:18:13 crc kubenswrapper[4824]: I1209 10:18:13.611515 4824 scope.go:117] "RemoveContainer" containerID="aa6386d2497fdcabbe806274e4229a9e73ef6471b24447710b88924b91553266" Dec 09 10:18:13 crc kubenswrapper[4824]: I1209 10:18:13.924456 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9d7c93f-e88e-4b55-91b8-f9dc9f42b912" path="/var/lib/kubelet/pods/d9d7c93f-e88e-4b55-91b8-f9dc9f42b912/volumes" Dec 09 10:18:14 crc kubenswrapper[4824]: I1209 10:18:14.464824 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"af7caa74-1882-4e75-9c52-942df2e40efe","Type":"ContainerStarted","Data":"dfc2cdea37c17df2242b9d0b5506979fb9210628638fd6e9ffb6ae0ed7c3008c"} Dec 09 10:18:14 crc kubenswrapper[4824]: I1209 10:18:14.477236 4824 generic.go:334] "Generic (PLEG): container finished" podID="c97a0c28-3a5a-430a-8000-e5c0518262b3" containerID="50c050207a10cb427846e70c07ba22db4dfae172195a3de54910a2b8b534fb61" exitCode=0 Dec 09 10:18:14 crc kubenswrapper[4824]: I1209 10:18:14.477297 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"c97a0c28-3a5a-430a-8000-e5c0518262b3","Type":"ContainerDied","Data":"50c050207a10cb427846e70c07ba22db4dfae172195a3de54910a2b8b534fb61"} Dec 09 10:18:15 crc kubenswrapper[4824]: I1209 10:18:15.514496 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"af7caa74-1882-4e75-9c52-942df2e40efe","Type":"ContainerStarted","Data":"eecb4d696781d7c777fbbc84f3a14a1e3f06dc01ba8901447c2014107b2feb65"} Dec 09 10:18:15 crc kubenswrapper[4824]: I1209 10:18:15.525177 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"c97a0c28-3a5a-430a-8000-e5c0518262b3","Type":"ContainerStarted","Data":"e25b15dbd243837fbbd7e6ba5757669059bf67a61c19bf3487c7bc9e455534af"} Dec 09 10:18:15 crc kubenswrapper[4824]: I1209 10:18:15.527959 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-2" Dec 09 10:18:15 crc kubenswrapper[4824]: I1209 10:18:15.561860 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-2" podStartSLOduration=37.561835697 podStartE2EDuration="37.561835697s" podCreationTimestamp="2025-12-09 10:17:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:18:15.554433313 +0000 UTC m=+1851.888938000" watchObservedRunningTime="2025-12-09 10:18:15.561835697 +0000 UTC m=+1851.896340374" Dec 09 10:18:16 crc kubenswrapper[4824]: I1209 10:18:16.550228 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"af7caa74-1882-4e75-9c52-942df2e40efe","Type":"ContainerStarted","Data":"a078b36c6ef10432b90c2a7f5920f86989e91e008b42cb2fda8eff5c9f074ff3"} Dec 09 10:18:24 crc kubenswrapper[4824]: I1209 10:18:24.768424 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vbfns" event={"ID":"eeb8228d-77ed-45ad-b2dc-085a2ba8814e","Type":"ContainerStarted","Data":"4355a66ad30c4ad455048f176c53e2b33f37483f021628ce5f59616c5fee50f4"} Dec 09 10:18:24 crc kubenswrapper[4824]: I1209 10:18:24.773030 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"af7caa74-1882-4e75-9c52-942df2e40efe","Type":"ContainerStarted","Data":"b5e589c075faddb220a3fbb9daf4d0b9904215cf6291c1126e6c70c2fc66437c"} Dec 09 10:18:24 crc kubenswrapper[4824]: I1209 10:18:24.800643 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vbfns" podStartSLOduration=2.1040486449999998 podStartE2EDuration="16.800610965s" podCreationTimestamp="2025-12-09 10:18:08 +0000 UTC" firstStartedPulling="2025-12-09 10:18:09.249033737 +0000 UTC m=+1845.583538404" lastFinishedPulling="2025-12-09 10:18:23.945596057 +0000 UTC m=+1860.280100724" observedRunningTime="2025-12-09 10:18:24.786367586 +0000 UTC m=+1861.120872253" watchObservedRunningTime="2025-12-09 10:18:24.800610965 +0000 UTC m=+1861.135115632" Dec 09 10:18:24 crc kubenswrapper[4824]: I1209 10:18:24.817575 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=1.851286505 podStartE2EDuration="12.817546558s" podCreationTimestamp="2025-12-09 10:18:12 +0000 UTC" firstStartedPulling="2025-12-09 10:18:12.961610415 +0000 UTC m=+1849.296115092" lastFinishedPulling="2025-12-09 10:18:23.927870478 +0000 UTC m=+1860.262375145" observedRunningTime="2025-12-09 10:18:24.80552825 +0000 UTC m=+1861.140032917" watchObservedRunningTime="2025-12-09 10:18:24.817546558 +0000 UTC m=+1861.152051225" Dec 09 10:18:24 crc kubenswrapper[4824]: I1209 10:18:24.911844 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:18:24 crc kubenswrapper[4824]: E1209 10:18:24.912183 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:18:27 crc kubenswrapper[4824]: I1209 10:18:27.810837 4824 generic.go:334] "Generic (PLEG): container finished" podID="eeb8228d-77ed-45ad-b2dc-085a2ba8814e" containerID="4355a66ad30c4ad455048f176c53e2b33f37483f021628ce5f59616c5fee50f4" exitCode=0 Dec 09 10:18:27 crc kubenswrapper[4824]: I1209 10:18:27.810942 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vbfns" event={"ID":"eeb8228d-77ed-45ad-b2dc-085a2ba8814e","Type":"ContainerDied","Data":"4355a66ad30c4ad455048f176c53e2b33f37483f021628ce5f59616c5fee50f4"} Dec 09 10:18:29 crc kubenswrapper[4824]: I1209 10:18:29.434665 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vbfns" Dec 09 10:18:29 crc kubenswrapper[4824]: I1209 10:18:29.500827 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-2" Dec 09 10:18:29 crc kubenswrapper[4824]: I1209 10:18:29.571841 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-1"] Dec 09 10:18:29 crc kubenswrapper[4824]: I1209 10:18:29.614337 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kghr\" (UniqueName: \"kubernetes.io/projected/eeb8228d-77ed-45ad-b2dc-085a2ba8814e-kube-api-access-9kghr\") pod \"eeb8228d-77ed-45ad-b2dc-085a2ba8814e\" (UID: \"eeb8228d-77ed-45ad-b2dc-085a2ba8814e\") " Dec 09 10:18:29 crc kubenswrapper[4824]: I1209 10:18:29.614463 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eeb8228d-77ed-45ad-b2dc-085a2ba8814e-ssh-key\") pod \"eeb8228d-77ed-45ad-b2dc-085a2ba8814e\" (UID: \"eeb8228d-77ed-45ad-b2dc-085a2ba8814e\") " Dec 09 10:18:29 crc kubenswrapper[4824]: I1209 10:18:29.614663 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eeb8228d-77ed-45ad-b2dc-085a2ba8814e-inventory\") pod \"eeb8228d-77ed-45ad-b2dc-085a2ba8814e\" (UID: \"eeb8228d-77ed-45ad-b2dc-085a2ba8814e\") " Dec 09 10:18:29 crc kubenswrapper[4824]: I1209 10:18:29.625221 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eeb8228d-77ed-45ad-b2dc-085a2ba8814e-kube-api-access-9kghr" (OuterVolumeSpecName: "kube-api-access-9kghr") pod "eeb8228d-77ed-45ad-b2dc-085a2ba8814e" (UID: "eeb8228d-77ed-45ad-b2dc-085a2ba8814e"). InnerVolumeSpecName "kube-api-access-9kghr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:18:29 crc kubenswrapper[4824]: I1209 10:18:29.672743 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eeb8228d-77ed-45ad-b2dc-085a2ba8814e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "eeb8228d-77ed-45ad-b2dc-085a2ba8814e" (UID: "eeb8228d-77ed-45ad-b2dc-085a2ba8814e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:18:29 crc kubenswrapper[4824]: I1209 10:18:29.741410 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kghr\" (UniqueName: \"kubernetes.io/projected/eeb8228d-77ed-45ad-b2dc-085a2ba8814e-kube-api-access-9kghr\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:29 crc kubenswrapper[4824]: I1209 10:18:29.741442 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eeb8228d-77ed-45ad-b2dc-085a2ba8814e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:29 crc kubenswrapper[4824]: I1209 10:18:29.746747 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eeb8228d-77ed-45ad-b2dc-085a2ba8814e-inventory" (OuterVolumeSpecName: "inventory") pod "eeb8228d-77ed-45ad-b2dc-085a2ba8814e" (UID: "eeb8228d-77ed-45ad-b2dc-085a2ba8814e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:18:29 crc kubenswrapper[4824]: I1209 10:18:29.838932 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vbfns" event={"ID":"eeb8228d-77ed-45ad-b2dc-085a2ba8814e","Type":"ContainerDied","Data":"9a35c5c35cbb4abc8e5997e57c9a664d25ee72562df278069b12eecc5aa5d25f"} Dec 09 10:18:29 crc kubenswrapper[4824]: I1209 10:18:29.838975 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vbfns" Dec 09 10:18:29 crc kubenswrapper[4824]: I1209 10:18:29.838986 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a35c5c35cbb4abc8e5997e57c9a664d25ee72562df278069b12eecc5aa5d25f" Dec 09 10:18:29 crc kubenswrapper[4824]: I1209 10:18:29.844138 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eeb8228d-77ed-45ad-b2dc-085a2ba8814e-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:29 crc kubenswrapper[4824]: I1209 10:18:29.937431 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq"] Dec 09 10:18:29 crc kubenswrapper[4824]: E1209 10:18:29.938053 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eeb8228d-77ed-45ad-b2dc-085a2ba8814e" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 09 10:18:29 crc kubenswrapper[4824]: I1209 10:18:29.938072 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="eeb8228d-77ed-45ad-b2dc-085a2ba8814e" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 09 10:18:29 crc kubenswrapper[4824]: I1209 10:18:29.938352 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="eeb8228d-77ed-45ad-b2dc-085a2ba8814e" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 09 10:18:29 crc kubenswrapper[4824]: I1209 10:18:29.939373 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq" Dec 09 10:18:29 crc kubenswrapper[4824]: I1209 10:18:29.948928 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 10:18:29 crc kubenswrapper[4824]: I1209 10:18:29.949153 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 10:18:29 crc kubenswrapper[4824]: I1209 10:18:29.949283 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5nzld" Dec 09 10:18:29 crc kubenswrapper[4824]: I1209 10:18:29.949442 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 10:18:29 crc kubenswrapper[4824]: I1209 10:18:29.958923 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq"] Dec 09 10:18:30 crc kubenswrapper[4824]: I1209 10:18:30.049846 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88cab327-4ed1-492c-9fd1-11a1657ec22f-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq\" (UID: \"88cab327-4ed1-492c-9fd1-11a1657ec22f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq" Dec 09 10:18:30 crc kubenswrapper[4824]: I1209 10:18:30.049932 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88cab327-4ed1-492c-9fd1-11a1657ec22f-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq\" (UID: \"88cab327-4ed1-492c-9fd1-11a1657ec22f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq" Dec 09 10:18:30 crc kubenswrapper[4824]: I1209 10:18:30.050027 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88cab327-4ed1-492c-9fd1-11a1657ec22f-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq\" (UID: \"88cab327-4ed1-492c-9fd1-11a1657ec22f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq" Dec 09 10:18:30 crc kubenswrapper[4824]: I1209 10:18:30.050211 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9dx2\" (UniqueName: \"kubernetes.io/projected/88cab327-4ed1-492c-9fd1-11a1657ec22f-kube-api-access-g9dx2\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq\" (UID: \"88cab327-4ed1-492c-9fd1-11a1657ec22f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq" Dec 09 10:18:30 crc kubenswrapper[4824]: I1209 10:18:30.152772 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9dx2\" (UniqueName: \"kubernetes.io/projected/88cab327-4ed1-492c-9fd1-11a1657ec22f-kube-api-access-g9dx2\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq\" (UID: \"88cab327-4ed1-492c-9fd1-11a1657ec22f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq" Dec 09 10:18:30 crc kubenswrapper[4824]: I1209 10:18:30.153017 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88cab327-4ed1-492c-9fd1-11a1657ec22f-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq\" (UID: \"88cab327-4ed1-492c-9fd1-11a1657ec22f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq" Dec 09 10:18:30 crc kubenswrapper[4824]: I1209 10:18:30.153090 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88cab327-4ed1-492c-9fd1-11a1657ec22f-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq\" (UID: \"88cab327-4ed1-492c-9fd1-11a1657ec22f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq" Dec 09 10:18:30 crc kubenswrapper[4824]: I1209 10:18:30.153174 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88cab327-4ed1-492c-9fd1-11a1657ec22f-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq\" (UID: \"88cab327-4ed1-492c-9fd1-11a1657ec22f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq" Dec 09 10:18:30 crc kubenswrapper[4824]: I1209 10:18:30.178420 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88cab327-4ed1-492c-9fd1-11a1657ec22f-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq\" (UID: \"88cab327-4ed1-492c-9fd1-11a1657ec22f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq" Dec 09 10:18:30 crc kubenswrapper[4824]: I1209 10:18:30.178429 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88cab327-4ed1-492c-9fd1-11a1657ec22f-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq\" (UID: \"88cab327-4ed1-492c-9fd1-11a1657ec22f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq" Dec 09 10:18:30 crc kubenswrapper[4824]: I1209 10:18:30.178581 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88cab327-4ed1-492c-9fd1-11a1657ec22f-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq\" (UID: \"88cab327-4ed1-492c-9fd1-11a1657ec22f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq" Dec 09 10:18:30 crc kubenswrapper[4824]: I1209 10:18:30.189568 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9dx2\" (UniqueName: \"kubernetes.io/projected/88cab327-4ed1-492c-9fd1-11a1657ec22f-kube-api-access-g9dx2\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq\" (UID: \"88cab327-4ed1-492c-9fd1-11a1657ec22f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq" Dec 09 10:18:30 crc kubenswrapper[4824]: I1209 10:18:30.277635 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq" Dec 09 10:18:31 crc kubenswrapper[4824]: I1209 10:18:31.090253 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq"] Dec 09 10:18:31 crc kubenswrapper[4824]: I1209 10:18:31.901886 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq" event={"ID":"88cab327-4ed1-492c-9fd1-11a1657ec22f","Type":"ContainerStarted","Data":"5790927f6f1250d3a282a012acb9b8111c67f3ab1aa569db14f66ad5af4d04c1"} Dec 09 10:18:32 crc kubenswrapper[4824]: I1209 10:18:32.914208 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq" event={"ID":"88cab327-4ed1-492c-9fd1-11a1657ec22f","Type":"ContainerStarted","Data":"e3dbacf0be37ad7fe653fb783d5a5f75f56970ce07cd94c3108d9a84c0d17fc0"} Dec 09 10:18:32 crc kubenswrapper[4824]: I1209 10:18:32.950164 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq" podStartSLOduration=3.432053735 podStartE2EDuration="3.950133704s" podCreationTimestamp="2025-12-09 10:18:29 +0000 UTC" firstStartedPulling="2025-12-09 10:18:31.104331489 +0000 UTC m=+1867.438836156" lastFinishedPulling="2025-12-09 10:18:31.622411458 +0000 UTC m=+1867.956916125" observedRunningTime="2025-12-09 10:18:32.937477554 +0000 UTC m=+1869.271982241" watchObservedRunningTime="2025-12-09 10:18:32.950133704 +0000 UTC m=+1869.284638371" Dec 09 10:18:35 crc kubenswrapper[4824]: I1209 10:18:35.200070 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-1" podUID="cee7a7d4-090b-4db0-abf1-d3f861f8e594" containerName="rabbitmq" containerID="cri-o://a32db2372ee433ebf981f7b113fb4bd4943f354f7d1ee3d51e9058cc39f25000" gracePeriod=604795 Dec 09 10:18:35 crc kubenswrapper[4824]: I1209 10:18:35.840205 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-1" podUID="cee7a7d4-090b-4db0-abf1-d3f861f8e594" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.129:5671: connect: connection refused" Dec 09 10:18:35 crc kubenswrapper[4824]: I1209 10:18:35.910569 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:18:35 crc kubenswrapper[4824]: E1209 10:18:35.911287 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:18:41 crc kubenswrapper[4824]: I1209 10:18:41.922797 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.076650 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cee7a7d4-090b-4db0-abf1-d3f861f8e594-rabbitmq-plugins\") pod \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.077049 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cee7a7d4-090b-4db0-abf1-d3f861f8e594-rabbitmq-confd\") pod \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.077081 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cee7a7d4-090b-4db0-abf1-d3f861f8e594-rabbitmq-erlang-cookie\") pod \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.077102 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cee7a7d4-090b-4db0-abf1-d3f861f8e594-config-data\") pod \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.077125 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cee7a7d4-090b-4db0-abf1-d3f861f8e594-plugins-conf\") pod \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.077245 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cee7a7d4-090b-4db0-abf1-d3f861f8e594-rabbitmq-tls\") pod \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.077288 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cee7a7d4-090b-4db0-abf1-d3f861f8e594-erlang-cookie-secret\") pod \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.077873 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cee7a7d4-090b-4db0-abf1-d3f861f8e594-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "cee7a7d4-090b-4db0-abf1-d3f861f8e594" (UID: "cee7a7d4-090b-4db0-abf1-d3f861f8e594"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.078302 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cee7a7d4-090b-4db0-abf1-d3f861f8e594-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "cee7a7d4-090b-4db0-abf1-d3f861f8e594" (UID: "cee7a7d4-090b-4db0-abf1-d3f861f8e594"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.078457 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cee7a7d4-090b-4db0-abf1-d3f861f8e594-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "cee7a7d4-090b-4db0-abf1-d3f861f8e594" (UID: "cee7a7d4-090b-4db0-abf1-d3f861f8e594"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.081010 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0\") pod \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.081168 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cee7a7d4-090b-4db0-abf1-d3f861f8e594-pod-info\") pod \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.081215 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66mpx\" (UniqueName: \"kubernetes.io/projected/cee7a7d4-090b-4db0-abf1-d3f861f8e594-kube-api-access-66mpx\") pod \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.081262 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cee7a7d4-090b-4db0-abf1-d3f861f8e594-server-conf\") pod \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\" (UID: \"cee7a7d4-090b-4db0-abf1-d3f861f8e594\") " Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.082496 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cee7a7d4-090b-4db0-abf1-d3f861f8e594-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.082520 4824 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cee7a7d4-090b-4db0-abf1-d3f861f8e594-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.082530 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cee7a7d4-090b-4db0-abf1-d3f861f8e594-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.086311 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cee7a7d4-090b-4db0-abf1-d3f861f8e594-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "cee7a7d4-090b-4db0-abf1-d3f861f8e594" (UID: "cee7a7d4-090b-4db0-abf1-d3f861f8e594"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.091122 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cee7a7d4-090b-4db0-abf1-d3f861f8e594-kube-api-access-66mpx" (OuterVolumeSpecName: "kube-api-access-66mpx") pod "cee7a7d4-090b-4db0-abf1-d3f861f8e594" (UID: "cee7a7d4-090b-4db0-abf1-d3f861f8e594"). InnerVolumeSpecName "kube-api-access-66mpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.098851 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cee7a7d4-090b-4db0-abf1-d3f861f8e594-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "cee7a7d4-090b-4db0-abf1-d3f861f8e594" (UID: "cee7a7d4-090b-4db0-abf1-d3f861f8e594"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.107893 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/cee7a7d4-090b-4db0-abf1-d3f861f8e594-pod-info" (OuterVolumeSpecName: "pod-info") pod "cee7a7d4-090b-4db0-abf1-d3f861f8e594" (UID: "cee7a7d4-090b-4db0-abf1-d3f861f8e594"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.124582 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cee7a7d4-090b-4db0-abf1-d3f861f8e594-config-data" (OuterVolumeSpecName: "config-data") pod "cee7a7d4-090b-4db0-abf1-d3f861f8e594" (UID: "cee7a7d4-090b-4db0-abf1-d3f861f8e594"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.126721 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0" (OuterVolumeSpecName: "persistence") pod "cee7a7d4-090b-4db0-abf1-d3f861f8e594" (UID: "cee7a7d4-090b-4db0-abf1-d3f861f8e594"). InnerVolumeSpecName "pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.185402 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cee7a7d4-090b-4db0-abf1-d3f861f8e594-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.185445 4824 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cee7a7d4-090b-4db0-abf1-d3f861f8e594-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.185509 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0\") on node \"crc\" " Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.185527 4824 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cee7a7d4-090b-4db0-abf1-d3f861f8e594-pod-info\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.185540 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66mpx\" (UniqueName: \"kubernetes.io/projected/cee7a7d4-090b-4db0-abf1-d3f861f8e594-kube-api-access-66mpx\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.185553 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cee7a7d4-090b-4db0-abf1-d3f861f8e594-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.191004 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cee7a7d4-090b-4db0-abf1-d3f861f8e594-server-conf" (OuterVolumeSpecName: "server-conf") pod "cee7a7d4-090b-4db0-abf1-d3f861f8e594" (UID: "cee7a7d4-090b-4db0-abf1-d3f861f8e594"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.234524 4824 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.235150 4824 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0") on node "crc" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.262798 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cee7a7d4-090b-4db0-abf1-d3f861f8e594-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "cee7a7d4-090b-4db0-abf1-d3f861f8e594" (UID: "cee7a7d4-090b-4db0-abf1-d3f861f8e594"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.287450 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cee7a7d4-090b-4db0-abf1-d3f861f8e594-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.287661 4824 reconciler_common.go:293] "Volume detached for volume \"pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.287759 4824 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cee7a7d4-090b-4db0-abf1-d3f861f8e594-server-conf\") on node \"crc\" DevicePath \"\"" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.693163 4824 generic.go:334] "Generic (PLEG): container finished" podID="cee7a7d4-090b-4db0-abf1-d3f861f8e594" containerID="a32db2372ee433ebf981f7b113fb4bd4943f354f7d1ee3d51e9058cc39f25000" exitCode=0 Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.693266 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"cee7a7d4-090b-4db0-abf1-d3f861f8e594","Type":"ContainerDied","Data":"a32db2372ee433ebf981f7b113fb4bd4943f354f7d1ee3d51e9058cc39f25000"} Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.693329 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"cee7a7d4-090b-4db0-abf1-d3f861f8e594","Type":"ContainerDied","Data":"485b03fcc8ab59181a8886f62c62c507879b0e3ab6211d14df87de77b20ae512"} Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.693375 4824 scope.go:117] "RemoveContainer" containerID="a32db2372ee433ebf981f7b113fb4bd4943f354f7d1ee3d51e9058cc39f25000" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.693564 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.748282 4824 scope.go:117] "RemoveContainer" containerID="9340d5745e62248e475ec029542c6a7717fa8bb811727daef75c5eb26017e816" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.748468 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-1"] Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.787902 4824 scope.go:117] "RemoveContainer" containerID="a32db2372ee433ebf981f7b113fb4bd4943f354f7d1ee3d51e9058cc39f25000" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.788181 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-1"] Dec 09 10:18:42 crc kubenswrapper[4824]: E1209 10:18:42.794183 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a32db2372ee433ebf981f7b113fb4bd4943f354f7d1ee3d51e9058cc39f25000\": container with ID starting with a32db2372ee433ebf981f7b113fb4bd4943f354f7d1ee3d51e9058cc39f25000 not found: ID does not exist" containerID="a32db2372ee433ebf981f7b113fb4bd4943f354f7d1ee3d51e9058cc39f25000" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.794254 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a32db2372ee433ebf981f7b113fb4bd4943f354f7d1ee3d51e9058cc39f25000"} err="failed to get container status \"a32db2372ee433ebf981f7b113fb4bd4943f354f7d1ee3d51e9058cc39f25000\": rpc error: code = NotFound desc = could not find container \"a32db2372ee433ebf981f7b113fb4bd4943f354f7d1ee3d51e9058cc39f25000\": container with ID starting with a32db2372ee433ebf981f7b113fb4bd4943f354f7d1ee3d51e9058cc39f25000 not found: ID does not exist" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.794299 4824 scope.go:117] "RemoveContainer" containerID="9340d5745e62248e475ec029542c6a7717fa8bb811727daef75c5eb26017e816" Dec 09 10:18:42 crc kubenswrapper[4824]: E1209 10:18:42.794659 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9340d5745e62248e475ec029542c6a7717fa8bb811727daef75c5eb26017e816\": container with ID starting with 9340d5745e62248e475ec029542c6a7717fa8bb811727daef75c5eb26017e816 not found: ID does not exist" containerID="9340d5745e62248e475ec029542c6a7717fa8bb811727daef75c5eb26017e816" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.794689 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9340d5745e62248e475ec029542c6a7717fa8bb811727daef75c5eb26017e816"} err="failed to get container status \"9340d5745e62248e475ec029542c6a7717fa8bb811727daef75c5eb26017e816\": rpc error: code = NotFound desc = could not find container \"9340d5745e62248e475ec029542c6a7717fa8bb811727daef75c5eb26017e816\": container with ID starting with 9340d5745e62248e475ec029542c6a7717fa8bb811727daef75c5eb26017e816 not found: ID does not exist" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.821968 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-1"] Dec 09 10:18:42 crc kubenswrapper[4824]: E1209 10:18:42.822567 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cee7a7d4-090b-4db0-abf1-d3f861f8e594" containerName="setup-container" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.822581 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cee7a7d4-090b-4db0-abf1-d3f861f8e594" containerName="setup-container" Dec 09 10:18:42 crc kubenswrapper[4824]: E1209 10:18:42.822654 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cee7a7d4-090b-4db0-abf1-d3f861f8e594" containerName="rabbitmq" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.822661 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cee7a7d4-090b-4db0-abf1-d3f861f8e594" containerName="rabbitmq" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.822899 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="cee7a7d4-090b-4db0-abf1-d3f861f8e594" containerName="rabbitmq" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.824500 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.850613 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-1"] Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.961529 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5de00a63-1140-413d-be79-922acdae88bc-pod-info\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.961580 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5de00a63-1140-413d-be79-922acdae88bc-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.961633 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5de00a63-1140-413d-be79-922acdae88bc-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.961756 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5de00a63-1140-413d-be79-922acdae88bc-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.961904 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5de00a63-1140-413d-be79-922acdae88bc-server-conf\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.961968 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5de00a63-1140-413d-be79-922acdae88bc-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.962029 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5de00a63-1140-413d-be79-922acdae88bc-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.962369 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.962535 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvm8d\" (UniqueName: \"kubernetes.io/projected/5de00a63-1140-413d-be79-922acdae88bc-kube-api-access-gvm8d\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.962700 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5de00a63-1140-413d-be79-922acdae88bc-config-data\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:42 crc kubenswrapper[4824]: I1209 10:18:42.962830 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5de00a63-1140-413d-be79-922acdae88bc-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.065246 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvm8d\" (UniqueName: \"kubernetes.io/projected/5de00a63-1140-413d-be79-922acdae88bc-kube-api-access-gvm8d\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.065324 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5de00a63-1140-413d-be79-922acdae88bc-config-data\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.065359 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5de00a63-1140-413d-be79-922acdae88bc-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.065448 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5de00a63-1140-413d-be79-922acdae88bc-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.065465 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5de00a63-1140-413d-be79-922acdae88bc-pod-info\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.065488 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5de00a63-1140-413d-be79-922acdae88bc-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.065503 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5de00a63-1140-413d-be79-922acdae88bc-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.065516 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5de00a63-1140-413d-be79-922acdae88bc-server-conf\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.065535 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5de00a63-1140-413d-be79-922acdae88bc-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.065568 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5de00a63-1140-413d-be79-922acdae88bc-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.065706 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.067848 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5de00a63-1140-413d-be79-922acdae88bc-config-data\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.068451 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5de00a63-1140-413d-be79-922acdae88bc-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.069036 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5de00a63-1140-413d-be79-922acdae88bc-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.069482 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5de00a63-1140-413d-be79-922acdae88bc-server-conf\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.069657 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5de00a63-1140-413d-be79-922acdae88bc-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.070878 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5de00a63-1140-413d-be79-922acdae88bc-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.071040 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.071070 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4f5f5da8e49a6f0edf98a2a4819436f9d70671dd23e36f231194d4ff6d8febbe/globalmount\"" pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.074476 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5de00a63-1140-413d-be79-922acdae88bc-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.075079 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5de00a63-1140-413d-be79-922acdae88bc-pod-info\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.075255 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5de00a63-1140-413d-be79-922acdae88bc-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.092038 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvm8d\" (UniqueName: \"kubernetes.io/projected/5de00a63-1140-413d-be79-922acdae88bc-kube-api-access-gvm8d\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.160874 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ca662358-f065-41e7-b9ab-b6d9429ddbc0\") pod \"rabbitmq-server-1\" (UID: \"5de00a63-1140-413d-be79-922acdae88bc\") " pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.470422 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Dec 09 10:18:43 crc kubenswrapper[4824]: I1209 10:18:43.957283 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cee7a7d4-090b-4db0-abf1-d3f861f8e594" path="/var/lib/kubelet/pods/cee7a7d4-090b-4db0-abf1-d3f861f8e594/volumes" Dec 09 10:18:44 crc kubenswrapper[4824]: I1209 10:18:44.291426 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-1"] Dec 09 10:18:44 crc kubenswrapper[4824]: I1209 10:18:44.935165 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"5de00a63-1140-413d-be79-922acdae88bc","Type":"ContainerStarted","Data":"63bb53815ecde67468e4f100b1d5d081468d5253108916fe1c1472f27d57b304"} Dec 09 10:18:47 crc kubenswrapper[4824]: I1209 10:18:47.910145 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:18:47 crc kubenswrapper[4824]: E1209 10:18:47.910862 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:18:52 crc kubenswrapper[4824]: I1209 10:18:52.164366 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"5de00a63-1140-413d-be79-922acdae88bc","Type":"ContainerStarted","Data":"64b26215638e542c1c7450492b22bb7a7aef90530055a1d7b21fabc90eeba4e6"} Dec 09 10:18:58 crc kubenswrapper[4824]: I1209 10:18:58.910439 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:18:58 crc kubenswrapper[4824]: E1209 10:18:58.912320 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:19:11 crc kubenswrapper[4824]: I1209 10:19:11.910581 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:19:11 crc kubenswrapper[4824]: E1209 10:19:11.911565 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:19:13 crc kubenswrapper[4824]: I1209 10:19:13.770122 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8rpnv"] Dec 09 10:19:13 crc kubenswrapper[4824]: I1209 10:19:13.773883 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rpnv" Dec 09 10:19:13 crc kubenswrapper[4824]: I1209 10:19:13.787868 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8rpnv"] Dec 09 10:19:13 crc kubenswrapper[4824]: I1209 10:19:13.826906 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhnfr\" (UniqueName: \"kubernetes.io/projected/5379faae-6dc0-4f93-a522-f192042b031e-kube-api-access-qhnfr\") pod \"certified-operators-8rpnv\" (UID: \"5379faae-6dc0-4f93-a522-f192042b031e\") " pod="openshift-marketplace/certified-operators-8rpnv" Dec 09 10:19:13 crc kubenswrapper[4824]: I1209 10:19:13.827323 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5379faae-6dc0-4f93-a522-f192042b031e-catalog-content\") pod \"certified-operators-8rpnv\" (UID: \"5379faae-6dc0-4f93-a522-f192042b031e\") " pod="openshift-marketplace/certified-operators-8rpnv" Dec 09 10:19:13 crc kubenswrapper[4824]: I1209 10:19:13.827505 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5379faae-6dc0-4f93-a522-f192042b031e-utilities\") pod \"certified-operators-8rpnv\" (UID: \"5379faae-6dc0-4f93-a522-f192042b031e\") " pod="openshift-marketplace/certified-operators-8rpnv" Dec 09 10:19:13 crc kubenswrapper[4824]: I1209 10:19:13.915205 4824 scope.go:117] "RemoveContainer" containerID="5b1cb02d98415e13ebd728dc83bdef09545646e7c9bbe6f2b829ceebe2c6ab0f" Dec 09 10:19:13 crc kubenswrapper[4824]: I1209 10:19:13.932051 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhnfr\" (UniqueName: \"kubernetes.io/projected/5379faae-6dc0-4f93-a522-f192042b031e-kube-api-access-qhnfr\") pod \"certified-operators-8rpnv\" (UID: \"5379faae-6dc0-4f93-a522-f192042b031e\") " pod="openshift-marketplace/certified-operators-8rpnv" Dec 09 10:19:13 crc kubenswrapper[4824]: I1209 10:19:13.932238 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5379faae-6dc0-4f93-a522-f192042b031e-catalog-content\") pod \"certified-operators-8rpnv\" (UID: \"5379faae-6dc0-4f93-a522-f192042b031e\") " pod="openshift-marketplace/certified-operators-8rpnv" Dec 09 10:19:13 crc kubenswrapper[4824]: I1209 10:19:13.932329 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5379faae-6dc0-4f93-a522-f192042b031e-utilities\") pod \"certified-operators-8rpnv\" (UID: \"5379faae-6dc0-4f93-a522-f192042b031e\") " pod="openshift-marketplace/certified-operators-8rpnv" Dec 09 10:19:13 crc kubenswrapper[4824]: I1209 10:19:13.934362 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5379faae-6dc0-4f93-a522-f192042b031e-catalog-content\") pod \"certified-operators-8rpnv\" (UID: \"5379faae-6dc0-4f93-a522-f192042b031e\") " pod="openshift-marketplace/certified-operators-8rpnv" Dec 09 10:19:13 crc kubenswrapper[4824]: I1209 10:19:13.935193 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5379faae-6dc0-4f93-a522-f192042b031e-utilities\") pod \"certified-operators-8rpnv\" (UID: \"5379faae-6dc0-4f93-a522-f192042b031e\") " pod="openshift-marketplace/certified-operators-8rpnv" Dec 09 10:19:13 crc kubenswrapper[4824]: I1209 10:19:13.962028 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhnfr\" (UniqueName: \"kubernetes.io/projected/5379faae-6dc0-4f93-a522-f192042b031e-kube-api-access-qhnfr\") pod \"certified-operators-8rpnv\" (UID: \"5379faae-6dc0-4f93-a522-f192042b031e\") " pod="openshift-marketplace/certified-operators-8rpnv" Dec 09 10:19:14 crc kubenswrapper[4824]: I1209 10:19:14.104111 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rpnv" Dec 09 10:19:15 crc kubenswrapper[4824]: I1209 10:19:15.120916 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8rpnv"] Dec 09 10:19:15 crc kubenswrapper[4824]: I1209 10:19:15.676988 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lk2f5"] Dec 09 10:19:15 crc kubenswrapper[4824]: I1209 10:19:15.686162 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lk2f5" Dec 09 10:19:15 crc kubenswrapper[4824]: I1209 10:19:15.692297 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lk2f5"] Dec 09 10:19:15 crc kubenswrapper[4824]: I1209 10:19:15.784696 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30da8ea3-fcf0-4978-ad49-30ff19f3708b-utilities\") pod \"redhat-marketplace-lk2f5\" (UID: \"30da8ea3-fcf0-4978-ad49-30ff19f3708b\") " pod="openshift-marketplace/redhat-marketplace-lk2f5" Dec 09 10:19:15 crc kubenswrapper[4824]: I1209 10:19:15.785085 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30da8ea3-fcf0-4978-ad49-30ff19f3708b-catalog-content\") pod \"redhat-marketplace-lk2f5\" (UID: \"30da8ea3-fcf0-4978-ad49-30ff19f3708b\") " pod="openshift-marketplace/redhat-marketplace-lk2f5" Dec 09 10:19:15 crc kubenswrapper[4824]: I1209 10:19:15.785219 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rzqq\" (UniqueName: \"kubernetes.io/projected/30da8ea3-fcf0-4978-ad49-30ff19f3708b-kube-api-access-4rzqq\") pod \"redhat-marketplace-lk2f5\" (UID: \"30da8ea3-fcf0-4978-ad49-30ff19f3708b\") " pod="openshift-marketplace/redhat-marketplace-lk2f5" Dec 09 10:19:15 crc kubenswrapper[4824]: I1209 10:19:15.876659 4824 generic.go:334] "Generic (PLEG): container finished" podID="5379faae-6dc0-4f93-a522-f192042b031e" containerID="846b46d91de59fc44f8b9b426daf3b29a6c8ffdb09a48a46d7c66c41bf97e9a0" exitCode=0 Dec 09 10:19:15 crc kubenswrapper[4824]: I1209 10:19:15.876776 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rpnv" event={"ID":"5379faae-6dc0-4f93-a522-f192042b031e","Type":"ContainerDied","Data":"846b46d91de59fc44f8b9b426daf3b29a6c8ffdb09a48a46d7c66c41bf97e9a0"} Dec 09 10:19:15 crc kubenswrapper[4824]: I1209 10:19:15.877040 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rpnv" event={"ID":"5379faae-6dc0-4f93-a522-f192042b031e","Type":"ContainerStarted","Data":"bdfd398f0c71b6c7f0fa741f848cd5a742d9e702f2c05b3008000049cdd8fccb"} Dec 09 10:19:15 crc kubenswrapper[4824]: I1209 10:19:15.889049 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30da8ea3-fcf0-4978-ad49-30ff19f3708b-utilities\") pod \"redhat-marketplace-lk2f5\" (UID: \"30da8ea3-fcf0-4978-ad49-30ff19f3708b\") " pod="openshift-marketplace/redhat-marketplace-lk2f5" Dec 09 10:19:15 crc kubenswrapper[4824]: I1209 10:19:15.889275 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30da8ea3-fcf0-4978-ad49-30ff19f3708b-catalog-content\") pod \"redhat-marketplace-lk2f5\" (UID: \"30da8ea3-fcf0-4978-ad49-30ff19f3708b\") " pod="openshift-marketplace/redhat-marketplace-lk2f5" Dec 09 10:19:15 crc kubenswrapper[4824]: I1209 10:19:15.889364 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rzqq\" (UniqueName: \"kubernetes.io/projected/30da8ea3-fcf0-4978-ad49-30ff19f3708b-kube-api-access-4rzqq\") pod \"redhat-marketplace-lk2f5\" (UID: \"30da8ea3-fcf0-4978-ad49-30ff19f3708b\") " pod="openshift-marketplace/redhat-marketplace-lk2f5" Dec 09 10:19:15 crc kubenswrapper[4824]: I1209 10:19:15.889700 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30da8ea3-fcf0-4978-ad49-30ff19f3708b-utilities\") pod \"redhat-marketplace-lk2f5\" (UID: \"30da8ea3-fcf0-4978-ad49-30ff19f3708b\") " pod="openshift-marketplace/redhat-marketplace-lk2f5" Dec 09 10:19:15 crc kubenswrapper[4824]: I1209 10:19:15.889826 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30da8ea3-fcf0-4978-ad49-30ff19f3708b-catalog-content\") pod \"redhat-marketplace-lk2f5\" (UID: \"30da8ea3-fcf0-4978-ad49-30ff19f3708b\") " pod="openshift-marketplace/redhat-marketplace-lk2f5" Dec 09 10:19:15 crc kubenswrapper[4824]: I1209 10:19:15.939194 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rzqq\" (UniqueName: \"kubernetes.io/projected/30da8ea3-fcf0-4978-ad49-30ff19f3708b-kube-api-access-4rzqq\") pod \"redhat-marketplace-lk2f5\" (UID: \"30da8ea3-fcf0-4978-ad49-30ff19f3708b\") " pod="openshift-marketplace/redhat-marketplace-lk2f5" Dec 09 10:19:16 crc kubenswrapper[4824]: I1209 10:19:16.019237 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lk2f5" Dec 09 10:19:16 crc kubenswrapper[4824]: I1209 10:19:16.813968 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lk2f5"] Dec 09 10:19:16 crc kubenswrapper[4824]: I1209 10:19:16.891129 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lk2f5" event={"ID":"30da8ea3-fcf0-4978-ad49-30ff19f3708b","Type":"ContainerStarted","Data":"66465ebbb30f2893c0e62b81706671fa925af4cd7f019cb1a5119ccd83edcb1f"} Dec 09 10:19:17 crc kubenswrapper[4824]: I1209 10:19:17.906447 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rpnv" event={"ID":"5379faae-6dc0-4f93-a522-f192042b031e","Type":"ContainerStarted","Data":"90c5bde9f293049cc4b5f70d721be6c2f5e5ad733d1bb09d52e2aa86efd6a54a"} Dec 09 10:19:17 crc kubenswrapper[4824]: I1209 10:19:17.908702 4824 generic.go:334] "Generic (PLEG): container finished" podID="30da8ea3-fcf0-4978-ad49-30ff19f3708b" containerID="a1d4fe78ef47441b0bf2f873f312305350c1de4b280cedc8a67a485eb026196d" exitCode=0 Dec 09 10:19:17 crc kubenswrapper[4824]: I1209 10:19:17.908758 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lk2f5" event={"ID":"30da8ea3-fcf0-4978-ad49-30ff19f3708b","Type":"ContainerDied","Data":"a1d4fe78ef47441b0bf2f873f312305350c1de4b280cedc8a67a485eb026196d"} Dec 09 10:19:20 crc kubenswrapper[4824]: I1209 10:19:20.199775 4824 generic.go:334] "Generic (PLEG): container finished" podID="5379faae-6dc0-4f93-a522-f192042b031e" containerID="90c5bde9f293049cc4b5f70d721be6c2f5e5ad733d1bb09d52e2aa86efd6a54a" exitCode=0 Dec 09 10:19:20 crc kubenswrapper[4824]: I1209 10:19:20.200032 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rpnv" event={"ID":"5379faae-6dc0-4f93-a522-f192042b031e","Type":"ContainerDied","Data":"90c5bde9f293049cc4b5f70d721be6c2f5e5ad733d1bb09d52e2aa86efd6a54a"} Dec 09 10:19:20 crc kubenswrapper[4824]: I1209 10:19:20.218011 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lk2f5" event={"ID":"30da8ea3-fcf0-4978-ad49-30ff19f3708b","Type":"ContainerStarted","Data":"0818129f196c991306877a074edffc519e004db062032539bc8fe40d9c10aca3"} Dec 09 10:19:21 crc kubenswrapper[4824]: I1209 10:19:21.229524 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rpnv" event={"ID":"5379faae-6dc0-4f93-a522-f192042b031e","Type":"ContainerStarted","Data":"80ed2a1496ad523b0847896ca4ffbd71d175cdce2b60011404f4ace2ff16dc61"} Dec 09 10:19:21 crc kubenswrapper[4824]: I1209 10:19:21.232254 4824 generic.go:334] "Generic (PLEG): container finished" podID="30da8ea3-fcf0-4978-ad49-30ff19f3708b" containerID="0818129f196c991306877a074edffc519e004db062032539bc8fe40d9c10aca3" exitCode=0 Dec 09 10:19:21 crc kubenswrapper[4824]: I1209 10:19:21.232304 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lk2f5" event={"ID":"30da8ea3-fcf0-4978-ad49-30ff19f3708b","Type":"ContainerDied","Data":"0818129f196c991306877a074edffc519e004db062032539bc8fe40d9c10aca3"} Dec 09 10:19:21 crc kubenswrapper[4824]: I1209 10:19:21.265278 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8rpnv" podStartSLOduration=3.403386826 podStartE2EDuration="8.265238527s" podCreationTimestamp="2025-12-09 10:19:13 +0000 UTC" firstStartedPulling="2025-12-09 10:19:15.879715261 +0000 UTC m=+1912.214219928" lastFinishedPulling="2025-12-09 10:19:20.741566922 +0000 UTC m=+1917.076071629" observedRunningTime="2025-12-09 10:19:21.252099482 +0000 UTC m=+1917.586604169" watchObservedRunningTime="2025-12-09 10:19:21.265238527 +0000 UTC m=+1917.599743184" Dec 09 10:19:22 crc kubenswrapper[4824]: I1209 10:19:22.251386 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lk2f5" event={"ID":"30da8ea3-fcf0-4978-ad49-30ff19f3708b","Type":"ContainerStarted","Data":"26537d670a9b0b49ff75051fa42881905f580b7c4f4a94e8c244e62ecb4eb318"} Dec 09 10:19:22 crc kubenswrapper[4824]: I1209 10:19:22.287073 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lk2f5" podStartSLOduration=3.570171767 podStartE2EDuration="7.287050911s" podCreationTimestamp="2025-12-09 10:19:15 +0000 UTC" firstStartedPulling="2025-12-09 10:19:17.910712132 +0000 UTC m=+1914.245216799" lastFinishedPulling="2025-12-09 10:19:21.627591276 +0000 UTC m=+1917.962095943" observedRunningTime="2025-12-09 10:19:22.275572859 +0000 UTC m=+1918.610077536" watchObservedRunningTime="2025-12-09 10:19:22.287050911 +0000 UTC m=+1918.621555578" Dec 09 10:19:24 crc kubenswrapper[4824]: I1209 10:19:24.104380 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8rpnv" Dec 09 10:19:24 crc kubenswrapper[4824]: I1209 10:19:24.104973 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8rpnv" Dec 09 10:19:24 crc kubenswrapper[4824]: I1209 10:19:24.156229 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8rpnv" Dec 09 10:19:24 crc kubenswrapper[4824]: I1209 10:19:24.911003 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:19:24 crc kubenswrapper[4824]: E1209 10:19:24.911836 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:19:26 crc kubenswrapper[4824]: I1209 10:19:26.020602 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lk2f5" Dec 09 10:19:26 crc kubenswrapper[4824]: I1209 10:19:26.020978 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lk2f5" Dec 09 10:19:26 crc kubenswrapper[4824]: I1209 10:19:26.070364 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lk2f5" Dec 09 10:19:26 crc kubenswrapper[4824]: I1209 10:19:26.357228 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lk2f5" Dec 09 10:19:27 crc kubenswrapper[4824]: I1209 10:19:27.670096 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lk2f5"] Dec 09 10:19:28 crc kubenswrapper[4824]: I1209 10:19:28.321575 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lk2f5" podUID="30da8ea3-fcf0-4978-ad49-30ff19f3708b" containerName="registry-server" containerID="cri-o://26537d670a9b0b49ff75051fa42881905f580b7c4f4a94e8c244e62ecb4eb318" gracePeriod=2 Dec 09 10:19:28 crc kubenswrapper[4824]: I1209 10:19:28.850665 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lk2f5" Dec 09 10:19:28 crc kubenswrapper[4824]: I1209 10:19:28.898907 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30da8ea3-fcf0-4978-ad49-30ff19f3708b-catalog-content\") pod \"30da8ea3-fcf0-4978-ad49-30ff19f3708b\" (UID: \"30da8ea3-fcf0-4978-ad49-30ff19f3708b\") " Dec 09 10:19:28 crc kubenswrapper[4824]: I1209 10:19:28.898955 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rzqq\" (UniqueName: \"kubernetes.io/projected/30da8ea3-fcf0-4978-ad49-30ff19f3708b-kube-api-access-4rzqq\") pod \"30da8ea3-fcf0-4978-ad49-30ff19f3708b\" (UID: \"30da8ea3-fcf0-4978-ad49-30ff19f3708b\") " Dec 09 10:19:28 crc kubenswrapper[4824]: I1209 10:19:28.899264 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30da8ea3-fcf0-4978-ad49-30ff19f3708b-utilities\") pod \"30da8ea3-fcf0-4978-ad49-30ff19f3708b\" (UID: \"30da8ea3-fcf0-4978-ad49-30ff19f3708b\") " Dec 09 10:19:28 crc kubenswrapper[4824]: I1209 10:19:28.901360 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30da8ea3-fcf0-4978-ad49-30ff19f3708b-utilities" (OuterVolumeSpecName: "utilities") pod "30da8ea3-fcf0-4978-ad49-30ff19f3708b" (UID: "30da8ea3-fcf0-4978-ad49-30ff19f3708b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:19:28 crc kubenswrapper[4824]: I1209 10:19:28.910504 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30da8ea3-fcf0-4978-ad49-30ff19f3708b-kube-api-access-4rzqq" (OuterVolumeSpecName: "kube-api-access-4rzqq") pod "30da8ea3-fcf0-4978-ad49-30ff19f3708b" (UID: "30da8ea3-fcf0-4978-ad49-30ff19f3708b"). InnerVolumeSpecName "kube-api-access-4rzqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:19:28 crc kubenswrapper[4824]: I1209 10:19:28.921154 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30da8ea3-fcf0-4978-ad49-30ff19f3708b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30da8ea3-fcf0-4978-ad49-30ff19f3708b" (UID: "30da8ea3-fcf0-4978-ad49-30ff19f3708b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:19:29 crc kubenswrapper[4824]: I1209 10:19:29.002602 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30da8ea3-fcf0-4978-ad49-30ff19f3708b-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 10:19:29 crc kubenswrapper[4824]: I1209 10:19:29.002636 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30da8ea3-fcf0-4978-ad49-30ff19f3708b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 10:19:29 crc kubenswrapper[4824]: I1209 10:19:29.002649 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rzqq\" (UniqueName: \"kubernetes.io/projected/30da8ea3-fcf0-4978-ad49-30ff19f3708b-kube-api-access-4rzqq\") on node \"crc\" DevicePath \"\"" Dec 09 10:19:29 crc kubenswrapper[4824]: I1209 10:19:29.338940 4824 generic.go:334] "Generic (PLEG): container finished" podID="30da8ea3-fcf0-4978-ad49-30ff19f3708b" containerID="26537d670a9b0b49ff75051fa42881905f580b7c4f4a94e8c244e62ecb4eb318" exitCode=0 Dec 09 10:19:29 crc kubenswrapper[4824]: I1209 10:19:29.339017 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lk2f5" Dec 09 10:19:29 crc kubenswrapper[4824]: I1209 10:19:29.339029 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lk2f5" event={"ID":"30da8ea3-fcf0-4978-ad49-30ff19f3708b","Type":"ContainerDied","Data":"26537d670a9b0b49ff75051fa42881905f580b7c4f4a94e8c244e62ecb4eb318"} Dec 09 10:19:29 crc kubenswrapper[4824]: I1209 10:19:29.339121 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lk2f5" event={"ID":"30da8ea3-fcf0-4978-ad49-30ff19f3708b","Type":"ContainerDied","Data":"66465ebbb30f2893c0e62b81706671fa925af4cd7f019cb1a5119ccd83edcb1f"} Dec 09 10:19:29 crc kubenswrapper[4824]: I1209 10:19:29.339156 4824 scope.go:117] "RemoveContainer" containerID="26537d670a9b0b49ff75051fa42881905f580b7c4f4a94e8c244e62ecb4eb318" Dec 09 10:19:29 crc kubenswrapper[4824]: I1209 10:19:29.382883 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lk2f5"] Dec 09 10:19:29 crc kubenswrapper[4824]: I1209 10:19:29.383894 4824 scope.go:117] "RemoveContainer" containerID="0818129f196c991306877a074edffc519e004db062032539bc8fe40d9c10aca3" Dec 09 10:19:29 crc kubenswrapper[4824]: I1209 10:19:29.397140 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lk2f5"] Dec 09 10:19:29 crc kubenswrapper[4824]: I1209 10:19:29.417219 4824 scope.go:117] "RemoveContainer" containerID="a1d4fe78ef47441b0bf2f873f312305350c1de4b280cedc8a67a485eb026196d" Dec 09 10:19:29 crc kubenswrapper[4824]: I1209 10:19:29.483908 4824 scope.go:117] "RemoveContainer" containerID="26537d670a9b0b49ff75051fa42881905f580b7c4f4a94e8c244e62ecb4eb318" Dec 09 10:19:29 crc kubenswrapper[4824]: E1209 10:19:29.484352 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26537d670a9b0b49ff75051fa42881905f580b7c4f4a94e8c244e62ecb4eb318\": container with ID starting with 26537d670a9b0b49ff75051fa42881905f580b7c4f4a94e8c244e62ecb4eb318 not found: ID does not exist" containerID="26537d670a9b0b49ff75051fa42881905f580b7c4f4a94e8c244e62ecb4eb318" Dec 09 10:19:29 crc kubenswrapper[4824]: I1209 10:19:29.484419 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26537d670a9b0b49ff75051fa42881905f580b7c4f4a94e8c244e62ecb4eb318"} err="failed to get container status \"26537d670a9b0b49ff75051fa42881905f580b7c4f4a94e8c244e62ecb4eb318\": rpc error: code = NotFound desc = could not find container \"26537d670a9b0b49ff75051fa42881905f580b7c4f4a94e8c244e62ecb4eb318\": container with ID starting with 26537d670a9b0b49ff75051fa42881905f580b7c4f4a94e8c244e62ecb4eb318 not found: ID does not exist" Dec 09 10:19:29 crc kubenswrapper[4824]: I1209 10:19:29.484457 4824 scope.go:117] "RemoveContainer" containerID="0818129f196c991306877a074edffc519e004db062032539bc8fe40d9c10aca3" Dec 09 10:19:29 crc kubenswrapper[4824]: E1209 10:19:29.484979 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0818129f196c991306877a074edffc519e004db062032539bc8fe40d9c10aca3\": container with ID starting with 0818129f196c991306877a074edffc519e004db062032539bc8fe40d9c10aca3 not found: ID does not exist" containerID="0818129f196c991306877a074edffc519e004db062032539bc8fe40d9c10aca3" Dec 09 10:19:29 crc kubenswrapper[4824]: I1209 10:19:29.485015 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0818129f196c991306877a074edffc519e004db062032539bc8fe40d9c10aca3"} err="failed to get container status \"0818129f196c991306877a074edffc519e004db062032539bc8fe40d9c10aca3\": rpc error: code = NotFound desc = could not find container \"0818129f196c991306877a074edffc519e004db062032539bc8fe40d9c10aca3\": container with ID starting with 0818129f196c991306877a074edffc519e004db062032539bc8fe40d9c10aca3 not found: ID does not exist" Dec 09 10:19:29 crc kubenswrapper[4824]: I1209 10:19:29.485038 4824 scope.go:117] "RemoveContainer" containerID="a1d4fe78ef47441b0bf2f873f312305350c1de4b280cedc8a67a485eb026196d" Dec 09 10:19:29 crc kubenswrapper[4824]: E1209 10:19:29.485325 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1d4fe78ef47441b0bf2f873f312305350c1de4b280cedc8a67a485eb026196d\": container with ID starting with a1d4fe78ef47441b0bf2f873f312305350c1de4b280cedc8a67a485eb026196d not found: ID does not exist" containerID="a1d4fe78ef47441b0bf2f873f312305350c1de4b280cedc8a67a485eb026196d" Dec 09 10:19:29 crc kubenswrapper[4824]: I1209 10:19:29.485353 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1d4fe78ef47441b0bf2f873f312305350c1de4b280cedc8a67a485eb026196d"} err="failed to get container status \"a1d4fe78ef47441b0bf2f873f312305350c1de4b280cedc8a67a485eb026196d\": rpc error: code = NotFound desc = could not find container \"a1d4fe78ef47441b0bf2f873f312305350c1de4b280cedc8a67a485eb026196d\": container with ID starting with a1d4fe78ef47441b0bf2f873f312305350c1de4b280cedc8a67a485eb026196d not found: ID does not exist" Dec 09 10:19:29 crc kubenswrapper[4824]: I1209 10:19:29.925365 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30da8ea3-fcf0-4978-ad49-30ff19f3708b" path="/var/lib/kubelet/pods/30da8ea3-fcf0-4978-ad49-30ff19f3708b/volumes" Dec 09 10:19:34 crc kubenswrapper[4824]: I1209 10:19:34.157963 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8rpnv" Dec 09 10:19:34 crc kubenswrapper[4824]: I1209 10:19:34.312164 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8rpnv"] Dec 09 10:19:34 crc kubenswrapper[4824]: I1209 10:19:34.409917 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8rpnv" podUID="5379faae-6dc0-4f93-a522-f192042b031e" containerName="registry-server" containerID="cri-o://80ed2a1496ad523b0847896ca4ffbd71d175cdce2b60011404f4ace2ff16dc61" gracePeriod=2 Dec 09 10:19:34 crc kubenswrapper[4824]: E1209 10:19:34.566213 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5379faae_6dc0_4f93_a522_f192042b031e.slice/crio-conmon-80ed2a1496ad523b0847896ca4ffbd71d175cdce2b60011404f4ace2ff16dc61.scope\": RecentStats: unable to find data in memory cache]" Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.033606 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rpnv" Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.177155 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5379faae-6dc0-4f93-a522-f192042b031e-utilities\") pod \"5379faae-6dc0-4f93-a522-f192042b031e\" (UID: \"5379faae-6dc0-4f93-a522-f192042b031e\") " Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.177501 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhnfr\" (UniqueName: \"kubernetes.io/projected/5379faae-6dc0-4f93-a522-f192042b031e-kube-api-access-qhnfr\") pod \"5379faae-6dc0-4f93-a522-f192042b031e\" (UID: \"5379faae-6dc0-4f93-a522-f192042b031e\") " Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.177561 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5379faae-6dc0-4f93-a522-f192042b031e-catalog-content\") pod \"5379faae-6dc0-4f93-a522-f192042b031e\" (UID: \"5379faae-6dc0-4f93-a522-f192042b031e\") " Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.177825 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5379faae-6dc0-4f93-a522-f192042b031e-utilities" (OuterVolumeSpecName: "utilities") pod "5379faae-6dc0-4f93-a522-f192042b031e" (UID: "5379faae-6dc0-4f93-a522-f192042b031e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.183098 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5379faae-6dc0-4f93-a522-f192042b031e-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.187841 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5379faae-6dc0-4f93-a522-f192042b031e-kube-api-access-qhnfr" (OuterVolumeSpecName: "kube-api-access-qhnfr") pod "5379faae-6dc0-4f93-a522-f192042b031e" (UID: "5379faae-6dc0-4f93-a522-f192042b031e"). InnerVolumeSpecName "kube-api-access-qhnfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.236235 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5379faae-6dc0-4f93-a522-f192042b031e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5379faae-6dc0-4f93-a522-f192042b031e" (UID: "5379faae-6dc0-4f93-a522-f192042b031e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.285399 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhnfr\" (UniqueName: \"kubernetes.io/projected/5379faae-6dc0-4f93-a522-f192042b031e-kube-api-access-qhnfr\") on node \"crc\" DevicePath \"\"" Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.285443 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5379faae-6dc0-4f93-a522-f192042b031e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.427167 4824 generic.go:334] "Generic (PLEG): container finished" podID="5379faae-6dc0-4f93-a522-f192042b031e" containerID="80ed2a1496ad523b0847896ca4ffbd71d175cdce2b60011404f4ace2ff16dc61" exitCode=0 Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.427238 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rpnv" event={"ID":"5379faae-6dc0-4f93-a522-f192042b031e","Type":"ContainerDied","Data":"80ed2a1496ad523b0847896ca4ffbd71d175cdce2b60011404f4ace2ff16dc61"} Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.427266 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rpnv" event={"ID":"5379faae-6dc0-4f93-a522-f192042b031e","Type":"ContainerDied","Data":"bdfd398f0c71b6c7f0fa741f848cd5a742d9e702f2c05b3008000049cdd8fccb"} Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.427306 4824 scope.go:117] "RemoveContainer" containerID="80ed2a1496ad523b0847896ca4ffbd71d175cdce2b60011404f4ace2ff16dc61" Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.427259 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rpnv" Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.430168 4824 generic.go:334] "Generic (PLEG): container finished" podID="5de00a63-1140-413d-be79-922acdae88bc" containerID="64b26215638e542c1c7450492b22bb7a7aef90530055a1d7b21fabc90eeba4e6" exitCode=0 Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.430210 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"5de00a63-1140-413d-be79-922acdae88bc","Type":"ContainerDied","Data":"64b26215638e542c1c7450492b22bb7a7aef90530055a1d7b21fabc90eeba4e6"} Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.472135 4824 scope.go:117] "RemoveContainer" containerID="90c5bde9f293049cc4b5f70d721be6c2f5e5ad733d1bb09d52e2aa86efd6a54a" Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.550649 4824 scope.go:117] "RemoveContainer" containerID="846b46d91de59fc44f8b9b426daf3b29a6c8ffdb09a48a46d7c66c41bf97e9a0" Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.550964 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8rpnv"] Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.565432 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8rpnv"] Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.645564 4824 scope.go:117] "RemoveContainer" containerID="80ed2a1496ad523b0847896ca4ffbd71d175cdce2b60011404f4ace2ff16dc61" Dec 09 10:19:35 crc kubenswrapper[4824]: E1209 10:19:35.646147 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80ed2a1496ad523b0847896ca4ffbd71d175cdce2b60011404f4ace2ff16dc61\": container with ID starting with 80ed2a1496ad523b0847896ca4ffbd71d175cdce2b60011404f4ace2ff16dc61 not found: ID does not exist" containerID="80ed2a1496ad523b0847896ca4ffbd71d175cdce2b60011404f4ace2ff16dc61" Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.646192 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80ed2a1496ad523b0847896ca4ffbd71d175cdce2b60011404f4ace2ff16dc61"} err="failed to get container status \"80ed2a1496ad523b0847896ca4ffbd71d175cdce2b60011404f4ace2ff16dc61\": rpc error: code = NotFound desc = could not find container \"80ed2a1496ad523b0847896ca4ffbd71d175cdce2b60011404f4ace2ff16dc61\": container with ID starting with 80ed2a1496ad523b0847896ca4ffbd71d175cdce2b60011404f4ace2ff16dc61 not found: ID does not exist" Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.646223 4824 scope.go:117] "RemoveContainer" containerID="90c5bde9f293049cc4b5f70d721be6c2f5e5ad733d1bb09d52e2aa86efd6a54a" Dec 09 10:19:35 crc kubenswrapper[4824]: E1209 10:19:35.646641 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90c5bde9f293049cc4b5f70d721be6c2f5e5ad733d1bb09d52e2aa86efd6a54a\": container with ID starting with 90c5bde9f293049cc4b5f70d721be6c2f5e5ad733d1bb09d52e2aa86efd6a54a not found: ID does not exist" containerID="90c5bde9f293049cc4b5f70d721be6c2f5e5ad733d1bb09d52e2aa86efd6a54a" Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.646661 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90c5bde9f293049cc4b5f70d721be6c2f5e5ad733d1bb09d52e2aa86efd6a54a"} err="failed to get container status \"90c5bde9f293049cc4b5f70d721be6c2f5e5ad733d1bb09d52e2aa86efd6a54a\": rpc error: code = NotFound desc = could not find container \"90c5bde9f293049cc4b5f70d721be6c2f5e5ad733d1bb09d52e2aa86efd6a54a\": container with ID starting with 90c5bde9f293049cc4b5f70d721be6c2f5e5ad733d1bb09d52e2aa86efd6a54a not found: ID does not exist" Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.646677 4824 scope.go:117] "RemoveContainer" containerID="846b46d91de59fc44f8b9b426daf3b29a6c8ffdb09a48a46d7c66c41bf97e9a0" Dec 09 10:19:35 crc kubenswrapper[4824]: E1209 10:19:35.647028 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"846b46d91de59fc44f8b9b426daf3b29a6c8ffdb09a48a46d7c66c41bf97e9a0\": container with ID starting with 846b46d91de59fc44f8b9b426daf3b29a6c8ffdb09a48a46d7c66c41bf97e9a0 not found: ID does not exist" containerID="846b46d91de59fc44f8b9b426daf3b29a6c8ffdb09a48a46d7c66c41bf97e9a0" Dec 09 10:19:35 crc kubenswrapper[4824]: I1209 10:19:35.647067 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"846b46d91de59fc44f8b9b426daf3b29a6c8ffdb09a48a46d7c66c41bf97e9a0"} err="failed to get container status \"846b46d91de59fc44f8b9b426daf3b29a6c8ffdb09a48a46d7c66c41bf97e9a0\": rpc error: code = NotFound desc = could not find container \"846b46d91de59fc44f8b9b426daf3b29a6c8ffdb09a48a46d7c66c41bf97e9a0\": container with ID starting with 846b46d91de59fc44f8b9b426daf3b29a6c8ffdb09a48a46d7c66c41bf97e9a0 not found: ID does not exist" Dec 09 10:19:36 crc kubenswrapper[4824]: I1209 10:19:36.009669 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5379faae-6dc0-4f93-a522-f192042b031e" path="/var/lib/kubelet/pods/5379faae-6dc0-4f93-a522-f192042b031e/volumes" Dec 09 10:19:36 crc kubenswrapper[4824]: I1209 10:19:36.443070 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"5de00a63-1140-413d-be79-922acdae88bc","Type":"ContainerStarted","Data":"bd3119acfc41142eb6670814bb45525680f166fa68e8ace19795a8fea2e03849"} Dec 09 10:19:36 crc kubenswrapper[4824]: I1209 10:19:36.443354 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-1" Dec 09 10:19:36 crc kubenswrapper[4824]: I1209 10:19:36.477819 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-1" podStartSLOduration=54.477768731 podStartE2EDuration="54.477768731s" podCreationTimestamp="2025-12-09 10:18:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:19:36.467018131 +0000 UTC m=+1932.801522818" watchObservedRunningTime="2025-12-09 10:19:36.477768731 +0000 UTC m=+1932.812273398" Dec 09 10:19:39 crc kubenswrapper[4824]: I1209 10:19:39.910841 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:19:39 crc kubenswrapper[4824]: E1209 10:19:39.911435 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:19:40 crc kubenswrapper[4824]: I1209 10:19:40.048199 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-ccjw2"] Dec 09 10:19:40 crc kubenswrapper[4824]: I1209 10:19:40.061765 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-ccjw2"] Dec 09 10:19:40 crc kubenswrapper[4824]: I1209 10:19:40.074172 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-5vfks"] Dec 09 10:19:40 crc kubenswrapper[4824]: I1209 10:19:40.087535 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-5vfks"] Dec 09 10:19:41 crc kubenswrapper[4824]: I1209 10:19:41.925474 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02d954cc-c1d4-45c8-86a2-fe001a222616" path="/var/lib/kubelet/pods/02d954cc-c1d4-45c8-86a2-fe001a222616/volumes" Dec 09 10:19:41 crc kubenswrapper[4824]: I1209 10:19:41.927854 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="245fbd50-9e68-416a-a260-56363b243f55" path="/var/lib/kubelet/pods/245fbd50-9e68-416a-a260-56363b243f55/volumes" Dec 09 10:19:43 crc kubenswrapper[4824]: I1209 10:19:43.036535 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-7029-account-create-update-dkjjt"] Dec 09 10:19:43 crc kubenswrapper[4824]: I1209 10:19:43.053063 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-pztmz"] Dec 09 10:19:43 crc kubenswrapper[4824]: I1209 10:19:43.068570 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-pztmz"] Dec 09 10:19:43 crc kubenswrapper[4824]: I1209 10:19:43.082792 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-a94f-account-create-update-f2dbj"] Dec 09 10:19:43 crc kubenswrapper[4824]: I1209 10:19:43.098330 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-7029-account-create-update-dkjjt"] Dec 09 10:19:43 crc kubenswrapper[4824]: I1209 10:19:43.117134 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-a94f-account-create-update-f2dbj"] Dec 09 10:19:43 crc kubenswrapper[4824]: I1209 10:19:43.924412 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c6345d1-110d-4671-b27b-f8d2abcf4d4f" path="/var/lib/kubelet/pods/1c6345d1-110d-4671-b27b-f8d2abcf4d4f/volumes" Dec 09 10:19:43 crc kubenswrapper[4824]: I1209 10:19:43.925979 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b6a6023-0aa7-4778-be37-6de2222088c4" path="/var/lib/kubelet/pods/7b6a6023-0aa7-4778-be37-6de2222088c4/volumes" Dec 09 10:19:43 crc kubenswrapper[4824]: I1209 10:19:43.926947 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83cb8977-7594-4739-90ae-ba915822102e" path="/var/lib/kubelet/pods/83cb8977-7594-4739-90ae-ba915822102e/volumes" Dec 09 10:19:46 crc kubenswrapper[4824]: I1209 10:19:46.053386 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-b025-account-create-update-cxmdd"] Dec 09 10:19:46 crc kubenswrapper[4824]: I1209 10:19:46.070797 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-65e6-account-create-update-znnqb"] Dec 09 10:19:46 crc kubenswrapper[4824]: I1209 10:19:46.082118 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-fmfsh"] Dec 09 10:19:46 crc kubenswrapper[4824]: I1209 10:19:46.092592 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-b025-account-create-update-cxmdd"] Dec 09 10:19:46 crc kubenswrapper[4824]: I1209 10:19:46.103319 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-65e6-account-create-update-znnqb"] Dec 09 10:19:46 crc kubenswrapper[4824]: I1209 10:19:46.115265 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-fmfsh"] Dec 09 10:19:47 crc kubenswrapper[4824]: I1209 10:19:47.928346 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="033734aa-97c9-498f-8bae-992507b79f71" path="/var/lib/kubelet/pods/033734aa-97c9-498f-8bae-992507b79f71/volumes" Dec 09 10:19:47 crc kubenswrapper[4824]: I1209 10:19:47.929871 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ec0f64d-691b-4433-a92b-ae6ab0ae4971" path="/var/lib/kubelet/pods/1ec0f64d-691b-4433-a92b-ae6ab0ae4971/volumes" Dec 09 10:19:47 crc kubenswrapper[4824]: I1209 10:19:47.931389 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="373c7c35-5998-40d4-b40b-5024d81e5567" path="/var/lib/kubelet/pods/373c7c35-5998-40d4-b40b-5024d81e5567/volumes" Dec 09 10:19:53 crc kubenswrapper[4824]: I1209 10:19:53.474020 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-1" Dec 09 10:19:53 crc kubenswrapper[4824]: I1209 10:19:53.559490 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 10:19:54 crc kubenswrapper[4824]: I1209 10:19:54.911231 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:19:54 crc kubenswrapper[4824]: E1209 10:19:54.911751 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:19:55 crc kubenswrapper[4824]: I1209 10:19:55.058600 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-6387-account-create-update-62lwv"] Dec 09 10:19:55 crc kubenswrapper[4824]: I1209 10:19:55.082405 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-fwd2s"] Dec 09 10:19:55 crc kubenswrapper[4824]: I1209 10:19:55.094207 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-6387-account-create-update-62lwv"] Dec 09 10:19:55 crc kubenswrapper[4824]: I1209 10:19:55.108080 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-fwd2s"] Dec 09 10:19:55 crc kubenswrapper[4824]: I1209 10:19:55.926848 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="108f0080-5222-4e67-8cef-6911d8ae4ced" path="/var/lib/kubelet/pods/108f0080-5222-4e67-8cef-6911d8ae4ced/volumes" Dec 09 10:19:55 crc kubenswrapper[4824]: I1209 10:19:55.928161 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49a2f3f0-02e4-4ed8-93ed-04911062274a" path="/var/lib/kubelet/pods/49a2f3f0-02e4-4ed8-93ed-04911062274a/volumes" Dec 09 10:19:57 crc kubenswrapper[4824]: I1209 10:19:57.952009 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="30eb761a-7bff-43a5-adfb-40e90e17e4a6" containerName="rabbitmq" containerID="cri-o://9cc814f255a3e8d2d480b4a92ba0edbcbf0c02ed24e819d5e91db3f835e461ff" gracePeriod=604796 Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.843269 4824 generic.go:334] "Generic (PLEG): container finished" podID="30eb761a-7bff-43a5-adfb-40e90e17e4a6" containerID="9cc814f255a3e8d2d480b4a92ba0edbcbf0c02ed24e819d5e91db3f835e461ff" exitCode=0 Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.843383 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"30eb761a-7bff-43a5-adfb-40e90e17e4a6","Type":"ContainerDied","Data":"9cc814f255a3e8d2d480b4a92ba0edbcbf0c02ed24e819d5e91db3f835e461ff"} Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.844052 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"30eb761a-7bff-43a5-adfb-40e90e17e4a6","Type":"ContainerDied","Data":"fb119ab02c0c550da4f09eea1b5b56e3ca8d71a6d88b0573aceb4930df25f9d6"} Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.844069 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb119ab02c0c550da4f09eea1b5b56e3ca8d71a6d88b0573aceb4930df25f9d6" Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.864987 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.977124 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30eb761a-7bff-43a5-adfb-40e90e17e4a6-config-data\") pod \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.977206 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/30eb761a-7bff-43a5-adfb-40e90e17e4a6-rabbitmq-confd\") pod \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.977229 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/30eb761a-7bff-43a5-adfb-40e90e17e4a6-rabbitmq-erlang-cookie\") pod \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.977246 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/30eb761a-7bff-43a5-adfb-40e90e17e4a6-plugins-conf\") pod \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.977368 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/30eb761a-7bff-43a5-adfb-40e90e17e4a6-rabbitmq-tls\") pod \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.977395 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/30eb761a-7bff-43a5-adfb-40e90e17e4a6-pod-info\") pod \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.977436 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/30eb761a-7bff-43a5-adfb-40e90e17e4a6-server-conf\") pod \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.977466 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/30eb761a-7bff-43a5-adfb-40e90e17e4a6-rabbitmq-plugins\") pod \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.978814 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98\") pod \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.978896 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ct9gv\" (UniqueName: \"kubernetes.io/projected/30eb761a-7bff-43a5-adfb-40e90e17e4a6-kube-api-access-ct9gv\") pod \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.979040 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/30eb761a-7bff-43a5-adfb-40e90e17e4a6-erlang-cookie-secret\") pod \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\" (UID: \"30eb761a-7bff-43a5-adfb-40e90e17e4a6\") " Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.979381 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30eb761a-7bff-43a5-adfb-40e90e17e4a6-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "30eb761a-7bff-43a5-adfb-40e90e17e4a6" (UID: "30eb761a-7bff-43a5-adfb-40e90e17e4a6"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.980642 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30eb761a-7bff-43a5-adfb-40e90e17e4a6-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "30eb761a-7bff-43a5-adfb-40e90e17e4a6" (UID: "30eb761a-7bff-43a5-adfb-40e90e17e4a6"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.980699 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30eb761a-7bff-43a5-adfb-40e90e17e4a6-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "30eb761a-7bff-43a5-adfb-40e90e17e4a6" (UID: "30eb761a-7bff-43a5-adfb-40e90e17e4a6"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.981438 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/30eb761a-7bff-43a5-adfb-40e90e17e4a6-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.981484 4824 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/30eb761a-7bff-43a5-adfb-40e90e17e4a6-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.981499 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/30eb761a-7bff-43a5-adfb-40e90e17e4a6-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.986042 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30eb761a-7bff-43a5-adfb-40e90e17e4a6-kube-api-access-ct9gv" (OuterVolumeSpecName: "kube-api-access-ct9gv") pod "30eb761a-7bff-43a5-adfb-40e90e17e4a6" (UID: "30eb761a-7bff-43a5-adfb-40e90e17e4a6"). InnerVolumeSpecName "kube-api-access-ct9gv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.986291 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/30eb761a-7bff-43a5-adfb-40e90e17e4a6-pod-info" (OuterVolumeSpecName: "pod-info") pod "30eb761a-7bff-43a5-adfb-40e90e17e4a6" (UID: "30eb761a-7bff-43a5-adfb-40e90e17e4a6"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 09 10:20:05 crc kubenswrapper[4824]: I1209 10:20:05.997605 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30eb761a-7bff-43a5-adfb-40e90e17e4a6-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "30eb761a-7bff-43a5-adfb-40e90e17e4a6" (UID: "30eb761a-7bff-43a5-adfb-40e90e17e4a6"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.001119 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30eb761a-7bff-43a5-adfb-40e90e17e4a6-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "30eb761a-7bff-43a5-adfb-40e90e17e4a6" (UID: "30eb761a-7bff-43a5-adfb-40e90e17e4a6"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.036071 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98" (OuterVolumeSpecName: "persistence") pod "30eb761a-7bff-43a5-adfb-40e90e17e4a6" (UID: "30eb761a-7bff-43a5-adfb-40e90e17e4a6"). InnerVolumeSpecName "pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.053688 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30eb761a-7bff-43a5-adfb-40e90e17e4a6-config-data" (OuterVolumeSpecName: "config-data") pod "30eb761a-7bff-43a5-adfb-40e90e17e4a6" (UID: "30eb761a-7bff-43a5-adfb-40e90e17e4a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.080745 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30eb761a-7bff-43a5-adfb-40e90e17e4a6-server-conf" (OuterVolumeSpecName: "server-conf") pod "30eb761a-7bff-43a5-adfb-40e90e17e4a6" (UID: "30eb761a-7bff-43a5-adfb-40e90e17e4a6"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.082919 4824 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/30eb761a-7bff-43a5-adfb-40e90e17e4a6-server-conf\") on node \"crc\" DevicePath \"\"" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.083091 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98\") on node \"crc\" " Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.083174 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ct9gv\" (UniqueName: \"kubernetes.io/projected/30eb761a-7bff-43a5-adfb-40e90e17e4a6-kube-api-access-ct9gv\") on node \"crc\" DevicePath \"\"" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.083256 4824 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/30eb761a-7bff-43a5-adfb-40e90e17e4a6-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.083322 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30eb761a-7bff-43a5-adfb-40e90e17e4a6-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.083389 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/30eb761a-7bff-43a5-adfb-40e90e17e4a6-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.083446 4824 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/30eb761a-7bff-43a5-adfb-40e90e17e4a6-pod-info\") on node \"crc\" DevicePath \"\"" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.128910 4824 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.129107 4824 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98") on node "crc" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.187362 4824 reconciler_common.go:293] "Volume detached for volume \"pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98\") on node \"crc\" DevicePath \"\"" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.199388 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30eb761a-7bff-43a5-adfb-40e90e17e4a6-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "30eb761a-7bff-43a5-adfb-40e90e17e4a6" (UID: "30eb761a-7bff-43a5-adfb-40e90e17e4a6"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.289332 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/30eb761a-7bff-43a5-adfb-40e90e17e4a6-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.855453 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.899725 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.912330 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.948466 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 10:20:06 crc kubenswrapper[4824]: E1209 10:20:06.949172 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30da8ea3-fcf0-4978-ad49-30ff19f3708b" containerName="registry-server" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.949197 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="30da8ea3-fcf0-4978-ad49-30ff19f3708b" containerName="registry-server" Dec 09 10:20:06 crc kubenswrapper[4824]: E1209 10:20:06.949223 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30eb761a-7bff-43a5-adfb-40e90e17e4a6" containerName="setup-container" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.949230 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="30eb761a-7bff-43a5-adfb-40e90e17e4a6" containerName="setup-container" Dec 09 10:20:06 crc kubenswrapper[4824]: E1209 10:20:06.949245 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5379faae-6dc0-4f93-a522-f192042b031e" containerName="extract-utilities" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.949252 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5379faae-6dc0-4f93-a522-f192042b031e" containerName="extract-utilities" Dec 09 10:20:06 crc kubenswrapper[4824]: E1209 10:20:06.949272 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5379faae-6dc0-4f93-a522-f192042b031e" containerName="extract-content" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.949278 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5379faae-6dc0-4f93-a522-f192042b031e" containerName="extract-content" Dec 09 10:20:06 crc kubenswrapper[4824]: E1209 10:20:06.949485 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5379faae-6dc0-4f93-a522-f192042b031e" containerName="registry-server" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.949492 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5379faae-6dc0-4f93-a522-f192042b031e" containerName="registry-server" Dec 09 10:20:06 crc kubenswrapper[4824]: E1209 10:20:06.949509 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30da8ea3-fcf0-4978-ad49-30ff19f3708b" containerName="extract-utilities" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.949517 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="30da8ea3-fcf0-4978-ad49-30ff19f3708b" containerName="extract-utilities" Dec 09 10:20:06 crc kubenswrapper[4824]: E1209 10:20:06.949527 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30da8ea3-fcf0-4978-ad49-30ff19f3708b" containerName="extract-content" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.949533 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="30da8ea3-fcf0-4978-ad49-30ff19f3708b" containerName="extract-content" Dec 09 10:20:06 crc kubenswrapper[4824]: E1209 10:20:06.949550 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30eb761a-7bff-43a5-adfb-40e90e17e4a6" containerName="rabbitmq" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.949555 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="30eb761a-7bff-43a5-adfb-40e90e17e4a6" containerName="rabbitmq" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.949775 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5379faae-6dc0-4f93-a522-f192042b031e" containerName="registry-server" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.949830 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="30eb761a-7bff-43a5-adfb-40e90e17e4a6" containerName="rabbitmq" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.949851 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="30da8ea3-fcf0-4978-ad49-30ff19f3708b" containerName="registry-server" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.951211 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 09 10:20:06 crc kubenswrapper[4824]: I1209 10:20:06.976359 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.116050 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.116101 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/42e63d7a-a5a7-47e3-933f-8e868262c6d6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.116237 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/42e63d7a-a5a7-47e3-933f-8e868262c6d6-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.116307 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/42e63d7a-a5a7-47e3-933f-8e868262c6d6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.116372 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/42e63d7a-a5a7-47e3-933f-8e868262c6d6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.116455 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/42e63d7a-a5a7-47e3-933f-8e868262c6d6-config-data\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.116540 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/42e63d7a-a5a7-47e3-933f-8e868262c6d6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.117645 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/42e63d7a-a5a7-47e3-933f-8e868262c6d6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.117849 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/42e63d7a-a5a7-47e3-933f-8e868262c6d6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.117947 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsrf8\" (UniqueName: \"kubernetes.io/projected/42e63d7a-a5a7-47e3-933f-8e868262c6d6-kube-api-access-bsrf8\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.118017 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/42e63d7a-a5a7-47e3-933f-8e868262c6d6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.220434 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/42e63d7a-a5a7-47e3-933f-8e868262c6d6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.220516 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/42e63d7a-a5a7-47e3-933f-8e868262c6d6-config-data\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.220572 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/42e63d7a-a5a7-47e3-933f-8e868262c6d6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.220684 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/42e63d7a-a5a7-47e3-933f-8e868262c6d6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.220876 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/42e63d7a-a5a7-47e3-933f-8e868262c6d6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.220965 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsrf8\" (UniqueName: \"kubernetes.io/projected/42e63d7a-a5a7-47e3-933f-8e868262c6d6-kube-api-access-bsrf8\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.220995 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/42e63d7a-a5a7-47e3-933f-8e868262c6d6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.221091 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.221118 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/42e63d7a-a5a7-47e3-933f-8e868262c6d6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.221153 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/42e63d7a-a5a7-47e3-933f-8e868262c6d6-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.221183 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/42e63d7a-a5a7-47e3-933f-8e868262c6d6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.221194 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/42e63d7a-a5a7-47e3-933f-8e868262c6d6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.221409 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/42e63d7a-a5a7-47e3-933f-8e868262c6d6-config-data\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.221457 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/42e63d7a-a5a7-47e3-933f-8e868262c6d6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.221690 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/42e63d7a-a5a7-47e3-933f-8e868262c6d6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.222830 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/42e63d7a-a5a7-47e3-933f-8e868262c6d6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.230355 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/42e63d7a-a5a7-47e3-933f-8e868262c6d6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.232646 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/42e63d7a-a5a7-47e3-933f-8e868262c6d6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.234948 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/42e63d7a-a5a7-47e3-933f-8e868262c6d6-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.244671 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/42e63d7a-a5a7-47e3-933f-8e868262c6d6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.300134 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.300186 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/da6b281685d3b39c9f4d80b46538e4c0d69a86be120be6d56528f21843928b2b/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.317854 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsrf8\" (UniqueName: \"kubernetes.io/projected/42e63d7a-a5a7-47e3-933f-8e868262c6d6-kube-api-access-bsrf8\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.502630 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-88f82cd1-55f5-4e42-90f4-1e9cd74b2d98\") pod \"rabbitmq-server-0\" (UID: \"42e63d7a-a5a7-47e3-933f-8e868262c6d6\") " pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.579513 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.910807 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:20:07 crc kubenswrapper[4824]: E1209 10:20:07.911586 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:20:07 crc kubenswrapper[4824]: I1209 10:20:07.934602 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30eb761a-7bff-43a5-adfb-40e90e17e4a6" path="/var/lib/kubelet/pods/30eb761a-7bff-43a5-adfb-40e90e17e4a6/volumes" Dec 09 10:20:08 crc kubenswrapper[4824]: I1209 10:20:08.084301 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 10:20:08 crc kubenswrapper[4824]: I1209 10:20:08.880399 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"42e63d7a-a5a7-47e3-933f-8e868262c6d6","Type":"ContainerStarted","Data":"d68dd4d9021b1a48af296d1c1e780e93eac01ea795ccee5ff2d5bd90d78abd68"} Dec 09 10:20:10 crc kubenswrapper[4824]: I1209 10:20:10.907434 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"42e63d7a-a5a7-47e3-933f-8e868262c6d6","Type":"ContainerStarted","Data":"e69195cc09f4d12cdf9839e639d559da983ac8f4bd0739bd4ca9c5e861cf92ed"} Dec 09 10:20:14 crc kubenswrapper[4824]: I1209 10:20:14.020662 4824 scope.go:117] "RemoveContainer" containerID="50499ef0bdf5865315ab6b6f9e726f19a093a2cf234f18fb4b15b596a33a6ee9" Dec 09 10:20:14 crc kubenswrapper[4824]: I1209 10:20:14.054579 4824 scope.go:117] "RemoveContainer" containerID="ea6e20c5563c87a5c35987ea4d40a6bae863eea01a42d6e87ef1edb2fc2311b0" Dec 09 10:20:14 crc kubenswrapper[4824]: I1209 10:20:14.144132 4824 scope.go:117] "RemoveContainer" containerID="f5f9c6e449b6d579f129798f5c0a7d4f334e6eba1d5832bbd2c1e34b78142745" Dec 09 10:20:14 crc kubenswrapper[4824]: I1209 10:20:14.212237 4824 scope.go:117] "RemoveContainer" containerID="29daa08ae30f5804d16cd244260b909954b1a9f61e7b3822eaba0d7cc8642676" Dec 09 10:20:14 crc kubenswrapper[4824]: I1209 10:20:14.285233 4824 scope.go:117] "RemoveContainer" containerID="1478290470c9d624da9f9cfc256d4304430790b207cbe3f1255acb46080432f6" Dec 09 10:20:14 crc kubenswrapper[4824]: I1209 10:20:14.366432 4824 scope.go:117] "RemoveContainer" containerID="9cc814f255a3e8d2d480b4a92ba0edbcbf0c02ed24e819d5e91db3f835e461ff" Dec 09 10:20:14 crc kubenswrapper[4824]: I1209 10:20:14.443329 4824 scope.go:117] "RemoveContainer" containerID="17314388a85fcca06ec1542f5802fb055e6225086fc3e4296338e07364427259" Dec 09 10:20:14 crc kubenswrapper[4824]: I1209 10:20:14.474132 4824 scope.go:117] "RemoveContainer" containerID="6bb064295473fab572428ffb93bdfc3abbae6a6ba39d21f98cdb6d98d7a4700c" Dec 09 10:20:14 crc kubenswrapper[4824]: I1209 10:20:14.499879 4824 scope.go:117] "RemoveContainer" containerID="afa3e9b8e9f87b390281f45b6490b3d377461954b20ca3c6deaae6c025b61aa2" Dec 09 10:20:14 crc kubenswrapper[4824]: I1209 10:20:14.522834 4824 scope.go:117] "RemoveContainer" containerID="962c26d0568bf98785f55af084fd2975bb8811ab3e95b72ecb01db06de61f1c4" Dec 09 10:20:14 crc kubenswrapper[4824]: I1209 10:20:14.545671 4824 scope.go:117] "RemoveContainer" containerID="d56cf72eb935f510b885c872fa34865ff360c27b73d0ff6b22950b29e79b50e1" Dec 09 10:20:14 crc kubenswrapper[4824]: I1209 10:20:14.570253 4824 scope.go:117] "RemoveContainer" containerID="24701089b3049252c9f2d6a4217bcb6c2856b5c88e9859477a2fe57787dc088a" Dec 09 10:20:18 crc kubenswrapper[4824]: I1209 10:20:18.911291 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:20:18 crc kubenswrapper[4824]: E1209 10:20:18.912070 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:20:32 crc kubenswrapper[4824]: I1209 10:20:32.911968 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:20:33 crc kubenswrapper[4824]: I1209 10:20:33.069275 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-wbxqj"] Dec 09 10:20:33 crc kubenswrapper[4824]: I1209 10:20:33.086767 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-wbxqj"] Dec 09 10:20:33 crc kubenswrapper[4824]: I1209 10:20:33.927730 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80a3aebf-6b04-4e56-81cc-a632b53293b6" path="/var/lib/kubelet/pods/80a3aebf-6b04-4e56-81cc-a632b53293b6/volumes" Dec 09 10:20:34 crc kubenswrapper[4824]: I1209 10:20:34.224098 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerStarted","Data":"d4d2394221cb52bd0aff000b3ed8a1bb39ce26a5dee3baa44cc5e9dfefc07899"} Dec 09 10:20:43 crc kubenswrapper[4824]: I1209 10:20:43.405450 4824 generic.go:334] "Generic (PLEG): container finished" podID="42e63d7a-a5a7-47e3-933f-8e868262c6d6" containerID="e69195cc09f4d12cdf9839e639d559da983ac8f4bd0739bd4ca9c5e861cf92ed" exitCode=0 Dec 09 10:20:43 crc kubenswrapper[4824]: I1209 10:20:43.406834 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"42e63d7a-a5a7-47e3-933f-8e868262c6d6","Type":"ContainerDied","Data":"e69195cc09f4d12cdf9839e639d559da983ac8f4bd0739bd4ca9c5e861cf92ed"} Dec 09 10:20:44 crc kubenswrapper[4824]: I1209 10:20:44.422066 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"42e63d7a-a5a7-47e3-933f-8e868262c6d6","Type":"ContainerStarted","Data":"bf8e728c54d834452b481582065d342c395edc267b1476e0482cf62ff6b47633"} Dec 09 10:20:44 crc kubenswrapper[4824]: I1209 10:20:44.422571 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 09 10:20:44 crc kubenswrapper[4824]: I1209 10:20:44.453083 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.453064415 podStartE2EDuration="38.453064415s" podCreationTimestamp="2025-12-09 10:20:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:20:44.445954921 +0000 UTC m=+2000.780459588" watchObservedRunningTime="2025-12-09 10:20:44.453064415 +0000 UTC m=+2000.787569082" Dec 09 10:20:57 crc kubenswrapper[4824]: I1209 10:20:57.582998 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 09 10:21:03 crc kubenswrapper[4824]: I1209 10:21:03.048121 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-nxcqz"] Dec 09 10:21:03 crc kubenswrapper[4824]: I1209 10:21:03.062001 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-nxcqz"] Dec 09 10:21:04 crc kubenswrapper[4824]: I1209 10:21:04.111374 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ee07f30-f716-4d7a-b5dd-71de0e128d3f" path="/var/lib/kubelet/pods/7ee07f30-f716-4d7a-b5dd-71de0e128d3f/volumes" Dec 09 10:21:04 crc kubenswrapper[4824]: I1209 10:21:04.112992 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-zh7x5"] Dec 09 10:21:04 crc kubenswrapper[4824]: I1209 10:21:04.113023 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-zh7x5"] Dec 09 10:21:05 crc kubenswrapper[4824]: I1209 10:21:05.941828 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55e299a6-a633-4e32-9766-54cbff8053b0" path="/var/lib/kubelet/pods/55e299a6-a633-4e32-9766-54cbff8053b0/volumes" Dec 09 10:21:08 crc kubenswrapper[4824]: I1209 10:21:08.043811 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-719f-account-create-update-mlf5c"] Dec 09 10:21:08 crc kubenswrapper[4824]: I1209 10:21:08.063010 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-0196-account-create-update-llvmc"] Dec 09 10:21:08 crc kubenswrapper[4824]: I1209 10:21:08.078013 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-9mmtj"] Dec 09 10:21:08 crc kubenswrapper[4824]: I1209 10:21:08.094209 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-9px5r"] Dec 09 10:21:08 crc kubenswrapper[4824]: I1209 10:21:08.107578 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-719f-account-create-update-mlf5c"] Dec 09 10:21:08 crc kubenswrapper[4824]: I1209 10:21:08.121106 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-0196-account-create-update-llvmc"] Dec 09 10:21:08 crc kubenswrapper[4824]: I1209 10:21:08.130644 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-9mmtj"] Dec 09 10:21:08 crc kubenswrapper[4824]: I1209 10:21:08.141830 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-9px5r"] Dec 09 10:21:08 crc kubenswrapper[4824]: I1209 10:21:08.156504 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b08a-account-create-update-jx676"] Dec 09 10:21:08 crc kubenswrapper[4824]: I1209 10:21:08.169671 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-b08a-account-create-update-jx676"] Dec 09 10:21:08 crc kubenswrapper[4824]: I1209 10:21:08.183578 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-3340-account-create-update-fskfw"] Dec 09 10:21:08 crc kubenswrapper[4824]: I1209 10:21:08.196487 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-3340-account-create-update-fskfw"] Dec 09 10:21:09 crc kubenswrapper[4824]: I1209 10:21:09.926377 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48b5dcdc-81a1-49f9-b5b6-5c9445be07dc" path="/var/lib/kubelet/pods/48b5dcdc-81a1-49f9-b5b6-5c9445be07dc/volumes" Dec 09 10:21:09 crc kubenswrapper[4824]: I1209 10:21:09.928914 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55a79bdb-6f53-4f93-85cd-e674210cc7e7" path="/var/lib/kubelet/pods/55a79bdb-6f53-4f93-85cd-e674210cc7e7/volumes" Dec 09 10:21:09 crc kubenswrapper[4824]: I1209 10:21:09.930534 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="826695ab-42b3-496a-a67b-add8ab95d4c1" path="/var/lib/kubelet/pods/826695ab-42b3-496a-a67b-add8ab95d4c1/volumes" Dec 09 10:21:09 crc kubenswrapper[4824]: I1209 10:21:09.932823 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a2cda53-48d6-480c-9da8-1caec867aee5" path="/var/lib/kubelet/pods/8a2cda53-48d6-480c-9da8-1caec867aee5/volumes" Dec 09 10:21:09 crc kubenswrapper[4824]: I1209 10:21:09.935544 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae040f9f-f187-4477-9455-af94eda40a1a" path="/var/lib/kubelet/pods/ae040f9f-f187-4477-9455-af94eda40a1a/volumes" Dec 09 10:21:09 crc kubenswrapper[4824]: I1209 10:21:09.936368 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6a3773f-0af2-4536-a361-dd51cedd700f" path="/var/lib/kubelet/pods/f6a3773f-0af2-4536-a361-dd51cedd700f/volumes" Dec 09 10:21:13 crc kubenswrapper[4824]: I1209 10:21:13.043668 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-6swf9"] Dec 09 10:21:13 crc kubenswrapper[4824]: I1209 10:21:13.056951 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-6swf9"] Dec 09 10:21:13 crc kubenswrapper[4824]: I1209 10:21:13.939868 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da292326-c56e-42d0-bfe2-2b75b53c210d" path="/var/lib/kubelet/pods/da292326-c56e-42d0-bfe2-2b75b53c210d/volumes" Dec 09 10:21:14 crc kubenswrapper[4824]: I1209 10:21:14.844188 4824 scope.go:117] "RemoveContainer" containerID="3685b228d8569d658b8bb0a9ef88c43b270f4ad1d48e5b2abcacf4d6a2c84a3f" Dec 09 10:21:14 crc kubenswrapper[4824]: I1209 10:21:14.884629 4824 scope.go:117] "RemoveContainer" containerID="2d511d51e8516b43cd422d16759aa763e3a7165dc038bbd12bcb6f11d8cccb8f" Dec 09 10:21:14 crc kubenswrapper[4824]: I1209 10:21:14.943816 4824 scope.go:117] "RemoveContainer" containerID="83e42cbfb500fa046b54d883713a3e0fb3e759e84101124723e2810a0f44ecc7" Dec 09 10:21:15 crc kubenswrapper[4824]: I1209 10:21:15.005697 4824 scope.go:117] "RemoveContainer" containerID="671b4caf32e6e41880e92594a0ad626b69af708fa7f1ccd387e7fdd7f87aaa0f" Dec 09 10:21:15 crc kubenswrapper[4824]: I1209 10:21:15.032102 4824 scope.go:117] "RemoveContainer" containerID="7a681790e04916ca576ca683ac87cc0c9fdee0b5b4671a1988993676391802ec" Dec 09 10:21:15 crc kubenswrapper[4824]: I1209 10:21:15.068094 4824 scope.go:117] "RemoveContainer" containerID="f886bcd18faabc96f773b58bea74aa161fc7d6fa281648ab17854ca99895392f" Dec 09 10:21:15 crc kubenswrapper[4824]: I1209 10:21:15.114427 4824 scope.go:117] "RemoveContainer" containerID="445d9b2d240dbce5259e65a1dd08ca31ec4fe19d06a1709bf06377d6f083ebde" Dec 09 10:21:15 crc kubenswrapper[4824]: I1209 10:21:15.165010 4824 scope.go:117] "RemoveContainer" containerID="a5e2081631925849dc471ae1ff9d3dbd6d34197130fbfbbc666fc8fbd35e9fce" Dec 09 10:21:15 crc kubenswrapper[4824]: I1209 10:21:15.220368 4824 scope.go:117] "RemoveContainer" containerID="150ba3aafdd9b74a31610254498fb25cc5221404d539d30379abeb1ea5efa048" Dec 09 10:21:15 crc kubenswrapper[4824]: I1209 10:21:15.289023 4824 scope.go:117] "RemoveContainer" containerID="a6aae0cc2f2d73cc679c1e9ba8e70322c362a1997846b0fadb6d5a23a5230277" Dec 09 10:21:15 crc kubenswrapper[4824]: I1209 10:21:15.326770 4824 scope.go:117] "RemoveContainer" containerID="5fe76dc4f6bef87e4744e566fce2c7d463ed227e3cbca0a8f76da631b8063ee1" Dec 09 10:21:15 crc kubenswrapper[4824]: I1209 10:21:15.361937 4824 scope.go:117] "RemoveContainer" containerID="b96320be5c824b5fba1b470169895191d2b61ec0a5840ffe6704738b2b7bd816" Dec 09 10:21:15 crc kubenswrapper[4824]: I1209 10:21:15.392470 4824 scope.go:117] "RemoveContainer" containerID="a8d7ef59965802baa03a3321640fe5629f7313e2ac4298b53df52384cae5caa9" Dec 09 10:21:15 crc kubenswrapper[4824]: I1209 10:21:15.417219 4824 scope.go:117] "RemoveContainer" containerID="6d8a450a8b8d26e3ab8e5ca1f2315522285cd246059130be9689bb6ecf8274a7" Dec 09 10:21:15 crc kubenswrapper[4824]: I1209 10:21:15.443243 4824 scope.go:117] "RemoveContainer" containerID="6acbb8552ca41469eedb970660ce6259b53fc161ce19ea9368d90e4325235c02" Dec 09 10:21:15 crc kubenswrapper[4824]: I1209 10:21:15.470454 4824 scope.go:117] "RemoveContainer" containerID="d056fadb93420fa399bb35c0d9edfbda3533432b04327f7d6c526b8f3f20f64c" Dec 09 10:21:49 crc kubenswrapper[4824]: I1209 10:21:49.046021 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-mqjtv"] Dec 09 10:21:49 crc kubenswrapper[4824]: I1209 10:21:49.061688 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-mqjtv"] Dec 09 10:21:49 crc kubenswrapper[4824]: I1209 10:21:49.923718 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8eb90797-2763-4ffb-8fa5-32a1c8db5c84" path="/var/lib/kubelet/pods/8eb90797-2763-4ffb-8fa5-32a1c8db5c84/volumes" Dec 09 10:21:56 crc kubenswrapper[4824]: I1209 10:21:56.417165 4824 generic.go:334] "Generic (PLEG): container finished" podID="88cab327-4ed1-492c-9fd1-11a1657ec22f" containerID="e3dbacf0be37ad7fe653fb783d5a5f75f56970ce07cd94c3108d9a84c0d17fc0" exitCode=0 Dec 09 10:21:56 crc kubenswrapper[4824]: I1209 10:21:56.417668 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq" event={"ID":"88cab327-4ed1-492c-9fd1-11a1657ec22f","Type":"ContainerDied","Data":"e3dbacf0be37ad7fe653fb783d5a5f75f56970ce07cd94c3108d9a84c0d17fc0"} Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.081566 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.217152 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9dx2\" (UniqueName: \"kubernetes.io/projected/88cab327-4ed1-492c-9fd1-11a1657ec22f-kube-api-access-g9dx2\") pod \"88cab327-4ed1-492c-9fd1-11a1657ec22f\" (UID: \"88cab327-4ed1-492c-9fd1-11a1657ec22f\") " Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.217842 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88cab327-4ed1-492c-9fd1-11a1657ec22f-inventory\") pod \"88cab327-4ed1-492c-9fd1-11a1657ec22f\" (UID: \"88cab327-4ed1-492c-9fd1-11a1657ec22f\") " Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.217908 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88cab327-4ed1-492c-9fd1-11a1657ec22f-ssh-key\") pod \"88cab327-4ed1-492c-9fd1-11a1657ec22f\" (UID: \"88cab327-4ed1-492c-9fd1-11a1657ec22f\") " Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.217970 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88cab327-4ed1-492c-9fd1-11a1657ec22f-bootstrap-combined-ca-bundle\") pod \"88cab327-4ed1-492c-9fd1-11a1657ec22f\" (UID: \"88cab327-4ed1-492c-9fd1-11a1657ec22f\") " Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.394963 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88cab327-4ed1-492c-9fd1-11a1657ec22f-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "88cab327-4ed1-492c-9fd1-11a1657ec22f" (UID: "88cab327-4ed1-492c-9fd1-11a1657ec22f"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.396070 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88cab327-4ed1-492c-9fd1-11a1657ec22f-kube-api-access-g9dx2" (OuterVolumeSpecName: "kube-api-access-g9dx2") pod "88cab327-4ed1-492c-9fd1-11a1657ec22f" (UID: "88cab327-4ed1-492c-9fd1-11a1657ec22f"). InnerVolumeSpecName "kube-api-access-g9dx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.429923 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88cab327-4ed1-492c-9fd1-11a1657ec22f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "88cab327-4ed1-492c-9fd1-11a1657ec22f" (UID: "88cab327-4ed1-492c-9fd1-11a1657ec22f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.435117 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88cab327-4ed1-492c-9fd1-11a1657ec22f-inventory" (OuterVolumeSpecName: "inventory") pod "88cab327-4ed1-492c-9fd1-11a1657ec22f" (UID: "88cab327-4ed1-492c-9fd1-11a1657ec22f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.443260 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq" event={"ID":"88cab327-4ed1-492c-9fd1-11a1657ec22f","Type":"ContainerDied","Data":"5790927f6f1250d3a282a012acb9b8111c67f3ab1aa569db14f66ad5af4d04c1"} Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.443303 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5790927f6f1250d3a282a012acb9b8111c67f3ab1aa569db14f66ad5af4d04c1" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.445113 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.488854 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88cab327-4ed1-492c-9fd1-11a1657ec22f-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.488898 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88cab327-4ed1-492c-9fd1-11a1657ec22f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.488913 4824 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88cab327-4ed1-492c-9fd1-11a1657ec22f-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.488930 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9dx2\" (UniqueName: \"kubernetes.io/projected/88cab327-4ed1-492c-9fd1-11a1657ec22f-kube-api-access-g9dx2\") on node \"crc\" DevicePath \"\"" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.618007 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs"] Dec 09 10:21:58 crc kubenswrapper[4824]: E1209 10:21:58.618634 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88cab327-4ed1-492c-9fd1-11a1657ec22f" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.618651 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="88cab327-4ed1-492c-9fd1-11a1657ec22f" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.618913 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="88cab327-4ed1-492c-9fd1-11a1657ec22f" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.619862 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.622659 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.622911 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.623311 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5nzld" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.633739 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs"] Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.643722 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.696310 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwlvr\" (UniqueName: \"kubernetes.io/projected/4a2f92dc-7710-4e22-8e1a-d303580314c2-kube-api-access-vwlvr\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs\" (UID: \"4a2f92dc-7710-4e22-8e1a-d303580314c2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.696755 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a2f92dc-7710-4e22-8e1a-d303580314c2-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs\" (UID: \"4a2f92dc-7710-4e22-8e1a-d303580314c2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.697055 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a2f92dc-7710-4e22-8e1a-d303580314c2-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs\" (UID: \"4a2f92dc-7710-4e22-8e1a-d303580314c2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.799729 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwlvr\" (UniqueName: \"kubernetes.io/projected/4a2f92dc-7710-4e22-8e1a-d303580314c2-kube-api-access-vwlvr\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs\" (UID: \"4a2f92dc-7710-4e22-8e1a-d303580314c2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.800299 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a2f92dc-7710-4e22-8e1a-d303580314c2-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs\" (UID: \"4a2f92dc-7710-4e22-8e1a-d303580314c2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.800526 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a2f92dc-7710-4e22-8e1a-d303580314c2-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs\" (UID: \"4a2f92dc-7710-4e22-8e1a-d303580314c2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.804206 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a2f92dc-7710-4e22-8e1a-d303580314c2-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs\" (UID: \"4a2f92dc-7710-4e22-8e1a-d303580314c2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.805034 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a2f92dc-7710-4e22-8e1a-d303580314c2-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs\" (UID: \"4a2f92dc-7710-4e22-8e1a-d303580314c2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.825711 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwlvr\" (UniqueName: \"kubernetes.io/projected/4a2f92dc-7710-4e22-8e1a-d303580314c2-kube-api-access-vwlvr\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs\" (UID: \"4a2f92dc-7710-4e22-8e1a-d303580314c2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs" Dec 09 10:21:58 crc kubenswrapper[4824]: I1209 10:21:58.961483 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs" Dec 09 10:21:59 crc kubenswrapper[4824]: I1209 10:21:59.528152 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs"] Dec 09 10:22:00 crc kubenswrapper[4824]: I1209 10:22:00.470443 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs" event={"ID":"4a2f92dc-7710-4e22-8e1a-d303580314c2","Type":"ContainerStarted","Data":"bcba845388f43a065c6411df438419dc8a6000ec29e4cf23a7f22bc329a63294"} Dec 09 10:22:00 crc kubenswrapper[4824]: I1209 10:22:00.470811 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs" event={"ID":"4a2f92dc-7710-4e22-8e1a-d303580314c2","Type":"ContainerStarted","Data":"69f927d52786e74a048727d53815a0605b18cac2e07e01265d1d50ef62ca1ac7"} Dec 09 10:22:02 crc kubenswrapper[4824]: I1209 10:22:02.354858 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs" podStartSLOduration=3.7861570860000002 podStartE2EDuration="4.354828783s" podCreationTimestamp="2025-12-09 10:21:58 +0000 UTC" firstStartedPulling="2025-12-09 10:21:59.531157416 +0000 UTC m=+2075.865662083" lastFinishedPulling="2025-12-09 10:22:00.099829113 +0000 UTC m=+2076.434333780" observedRunningTime="2025-12-09 10:22:01.506236621 +0000 UTC m=+2077.840741328" watchObservedRunningTime="2025-12-09 10:22:02.354828783 +0000 UTC m=+2078.689333450" Dec 09 10:22:02 crc kubenswrapper[4824]: I1209 10:22:02.370411 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dmvc6"] Dec 09 10:22:02 crc kubenswrapper[4824]: I1209 10:22:02.373571 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dmvc6" Dec 09 10:22:02 crc kubenswrapper[4824]: I1209 10:22:02.391156 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dmvc6"] Dec 09 10:22:02 crc kubenswrapper[4824]: I1209 10:22:02.536433 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43d37775-34fa-4b0b-b56a-786ef557d7a9-utilities\") pod \"redhat-operators-dmvc6\" (UID: \"43d37775-34fa-4b0b-b56a-786ef557d7a9\") " pod="openshift-marketplace/redhat-operators-dmvc6" Dec 09 10:22:02 crc kubenswrapper[4824]: I1209 10:22:02.536552 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43d37775-34fa-4b0b-b56a-786ef557d7a9-catalog-content\") pod \"redhat-operators-dmvc6\" (UID: \"43d37775-34fa-4b0b-b56a-786ef557d7a9\") " pod="openshift-marketplace/redhat-operators-dmvc6" Dec 09 10:22:02 crc kubenswrapper[4824]: I1209 10:22:02.536672 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjqk7\" (UniqueName: \"kubernetes.io/projected/43d37775-34fa-4b0b-b56a-786ef557d7a9-kube-api-access-hjqk7\") pod \"redhat-operators-dmvc6\" (UID: \"43d37775-34fa-4b0b-b56a-786ef557d7a9\") " pod="openshift-marketplace/redhat-operators-dmvc6" Dec 09 10:22:02 crc kubenswrapper[4824]: I1209 10:22:02.639112 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43d37775-34fa-4b0b-b56a-786ef557d7a9-utilities\") pod \"redhat-operators-dmvc6\" (UID: \"43d37775-34fa-4b0b-b56a-786ef557d7a9\") " pod="openshift-marketplace/redhat-operators-dmvc6" Dec 09 10:22:02 crc kubenswrapper[4824]: I1209 10:22:02.640361 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43d37775-34fa-4b0b-b56a-786ef557d7a9-catalog-content\") pod \"redhat-operators-dmvc6\" (UID: \"43d37775-34fa-4b0b-b56a-786ef557d7a9\") " pod="openshift-marketplace/redhat-operators-dmvc6" Dec 09 10:22:02 crc kubenswrapper[4824]: I1209 10:22:02.639632 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43d37775-34fa-4b0b-b56a-786ef557d7a9-utilities\") pod \"redhat-operators-dmvc6\" (UID: \"43d37775-34fa-4b0b-b56a-786ef557d7a9\") " pod="openshift-marketplace/redhat-operators-dmvc6" Dec 09 10:22:02 crc kubenswrapper[4824]: I1209 10:22:02.640547 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjqk7\" (UniqueName: \"kubernetes.io/projected/43d37775-34fa-4b0b-b56a-786ef557d7a9-kube-api-access-hjqk7\") pod \"redhat-operators-dmvc6\" (UID: \"43d37775-34fa-4b0b-b56a-786ef557d7a9\") " pod="openshift-marketplace/redhat-operators-dmvc6" Dec 09 10:22:02 crc kubenswrapper[4824]: I1209 10:22:02.640832 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43d37775-34fa-4b0b-b56a-786ef557d7a9-catalog-content\") pod \"redhat-operators-dmvc6\" (UID: \"43d37775-34fa-4b0b-b56a-786ef557d7a9\") " pod="openshift-marketplace/redhat-operators-dmvc6" Dec 09 10:22:02 crc kubenswrapper[4824]: I1209 10:22:02.672562 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjqk7\" (UniqueName: \"kubernetes.io/projected/43d37775-34fa-4b0b-b56a-786ef557d7a9-kube-api-access-hjqk7\") pod \"redhat-operators-dmvc6\" (UID: \"43d37775-34fa-4b0b-b56a-786ef557d7a9\") " pod="openshift-marketplace/redhat-operators-dmvc6" Dec 09 10:22:02 crc kubenswrapper[4824]: I1209 10:22:02.710983 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dmvc6" Dec 09 10:22:03 crc kubenswrapper[4824]: I1209 10:22:03.347556 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dmvc6"] Dec 09 10:22:03 crc kubenswrapper[4824]: I1209 10:22:03.543222 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmvc6" event={"ID":"43d37775-34fa-4b0b-b56a-786ef557d7a9","Type":"ContainerStarted","Data":"7ba88db7a96c344a847f5509fd56bc6dc45203a2802708be6f72a671d2d383ff"} Dec 09 10:22:04 crc kubenswrapper[4824]: I1209 10:22:04.558524 4824 generic.go:334] "Generic (PLEG): container finished" podID="43d37775-34fa-4b0b-b56a-786ef557d7a9" containerID="23ae8f7deb3c111901743ab8b96b40f0f6d61167f8f406b24b13ce3fb3471697" exitCode=0 Dec 09 10:22:04 crc kubenswrapper[4824]: I1209 10:22:04.558877 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmvc6" event={"ID":"43d37775-34fa-4b0b-b56a-786ef557d7a9","Type":"ContainerDied","Data":"23ae8f7deb3c111901743ab8b96b40f0f6d61167f8f406b24b13ce3fb3471697"} Dec 09 10:22:05 crc kubenswrapper[4824]: I1209 10:22:05.573700 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmvc6" event={"ID":"43d37775-34fa-4b0b-b56a-786ef557d7a9","Type":"ContainerStarted","Data":"5e2aa93540e7a055512dc5e7b7b79eedd0aeb9c3f72004a9c28f848e059735d2"} Dec 09 10:22:08 crc kubenswrapper[4824]: I1209 10:22:08.147314 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-ql4lt"] Dec 09 10:22:08 crc kubenswrapper[4824]: I1209 10:22:08.180132 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-mlcsg"] Dec 09 10:22:08 crc kubenswrapper[4824]: I1209 10:22:08.218155 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-mlcsg"] Dec 09 10:22:08 crc kubenswrapper[4824]: I1209 10:22:08.248173 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-ql4lt"] Dec 09 10:22:08 crc kubenswrapper[4824]: I1209 10:22:08.261658 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-d4klr"] Dec 09 10:22:08 crc kubenswrapper[4824]: I1209 10:22:08.276946 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-d4klr"] Dec 09 10:22:09 crc kubenswrapper[4824]: I1209 10:22:09.635810 4824 generic.go:334] "Generic (PLEG): container finished" podID="43d37775-34fa-4b0b-b56a-786ef557d7a9" containerID="5e2aa93540e7a055512dc5e7b7b79eedd0aeb9c3f72004a9c28f848e059735d2" exitCode=0 Dec 09 10:22:09 crc kubenswrapper[4824]: I1209 10:22:09.635841 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmvc6" event={"ID":"43d37775-34fa-4b0b-b56a-786ef557d7a9","Type":"ContainerDied","Data":"5e2aa93540e7a055512dc5e7b7b79eedd0aeb9c3f72004a9c28f848e059735d2"} Dec 09 10:22:09 crc kubenswrapper[4824]: I1209 10:22:09.924177 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a07abd4-056e-4ea7-962f-60f5a03ea701" path="/var/lib/kubelet/pods/6a07abd4-056e-4ea7-962f-60f5a03ea701/volumes" Dec 09 10:22:09 crc kubenswrapper[4824]: I1209 10:22:09.925269 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aae4e0f1-0295-4964-8b34-56bdab58a152" path="/var/lib/kubelet/pods/aae4e0f1-0295-4964-8b34-56bdab58a152/volumes" Dec 09 10:22:09 crc kubenswrapper[4824]: I1209 10:22:09.926181 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3d076df-d1e1-473e-9584-79b30cb6d023" path="/var/lib/kubelet/pods/d3d076df-d1e1-473e-9584-79b30cb6d023/volumes" Dec 09 10:22:10 crc kubenswrapper[4824]: I1209 10:22:10.651495 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmvc6" event={"ID":"43d37775-34fa-4b0b-b56a-786ef557d7a9","Type":"ContainerStarted","Data":"5c7b1732a6d57c9bdc63b0f4d01d80f3bdc7dbc43fb7217b252ae5243e594470"} Dec 09 10:22:10 crc kubenswrapper[4824]: I1209 10:22:10.684852 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dmvc6" podStartSLOduration=3.199921125 podStartE2EDuration="8.684826598s" podCreationTimestamp="2025-12-09 10:22:02 +0000 UTC" firstStartedPulling="2025-12-09 10:22:04.561115616 +0000 UTC m=+2080.895620283" lastFinishedPulling="2025-12-09 10:22:10.046021089 +0000 UTC m=+2086.380525756" observedRunningTime="2025-12-09 10:22:10.673597244 +0000 UTC m=+2087.008101931" watchObservedRunningTime="2025-12-09 10:22:10.684826598 +0000 UTC m=+2087.019331265" Dec 09 10:22:12 crc kubenswrapper[4824]: I1209 10:22:12.711329 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dmvc6" Dec 09 10:22:12 crc kubenswrapper[4824]: I1209 10:22:12.711696 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dmvc6" Dec 09 10:22:13 crc kubenswrapper[4824]: I1209 10:22:13.778388 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dmvc6" podUID="43d37775-34fa-4b0b-b56a-786ef557d7a9" containerName="registry-server" probeResult="failure" output=< Dec 09 10:22:13 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 10:22:13 crc kubenswrapper[4824]: > Dec 09 10:22:15 crc kubenswrapper[4824]: I1209 10:22:15.775628 4824 scope.go:117] "RemoveContainer" containerID="c9d596aa7c77e733aff0fa6a7db404be9deaa883f07febf1dac4160a367c0c6e" Dec 09 10:22:15 crc kubenswrapper[4824]: I1209 10:22:15.815198 4824 scope.go:117] "RemoveContainer" containerID="b42b3ca64da6ae1b9dba2cc35d19d60ff830c508b053eef56824f66ff004c329" Dec 09 10:22:15 crc kubenswrapper[4824]: I1209 10:22:15.839161 4824 scope.go:117] "RemoveContainer" containerID="e215988c9459664b8ba3af9c71ac7573e5936c4f048333de2977d582be175243" Dec 09 10:22:15 crc kubenswrapper[4824]: I1209 10:22:15.883551 4824 scope.go:117] "RemoveContainer" containerID="31332a7c9ba75e6a67f37596510b470eb420a7ffcfbea609a3ac1efd24e7a3ec" Dec 09 10:22:15 crc kubenswrapper[4824]: I1209 10:22:15.950960 4824 scope.go:117] "RemoveContainer" containerID="481598142ccbbb37a8c1d4b3c9c41b36e6797b272a9cd42c11a24be6d054d4d6" Dec 09 10:22:16 crc kubenswrapper[4824]: I1209 10:22:16.015189 4824 scope.go:117] "RemoveContainer" containerID="3a153c578db1e11c3e8a67688b76a866e98ce0c6d3d6ff91aca0a32828d69f90" Dec 09 10:22:16 crc kubenswrapper[4824]: I1209 10:22:16.101098 4824 scope.go:117] "RemoveContainer" containerID="066750eead3e694d0dc6b7ddabcbde71a628f40e58f98c45727562ff2269f30c" Dec 09 10:22:16 crc kubenswrapper[4824]: I1209 10:22:16.134653 4824 scope.go:117] "RemoveContainer" containerID="66df44334c000ef2419a325e951be6d72b85973d4249b0d527f40d2ad0d6338b" Dec 09 10:22:22 crc kubenswrapper[4824]: I1209 10:22:22.787078 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dmvc6" Dec 09 10:22:22 crc kubenswrapper[4824]: I1209 10:22:22.857653 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dmvc6" Dec 09 10:22:23 crc kubenswrapper[4824]: I1209 10:22:23.056873 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dmvc6"] Dec 09 10:22:24 crc kubenswrapper[4824]: I1209 10:22:24.810710 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dmvc6" podUID="43d37775-34fa-4b0b-b56a-786ef557d7a9" containerName="registry-server" containerID="cri-o://5c7b1732a6d57c9bdc63b0f4d01d80f3bdc7dbc43fb7217b252ae5243e594470" gracePeriod=2 Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.448474 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dmvc6" Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.565917 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43d37775-34fa-4b0b-b56a-786ef557d7a9-catalog-content\") pod \"43d37775-34fa-4b0b-b56a-786ef557d7a9\" (UID: \"43d37775-34fa-4b0b-b56a-786ef557d7a9\") " Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.566195 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43d37775-34fa-4b0b-b56a-786ef557d7a9-utilities\") pod \"43d37775-34fa-4b0b-b56a-786ef557d7a9\" (UID: \"43d37775-34fa-4b0b-b56a-786ef557d7a9\") " Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.566279 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjqk7\" (UniqueName: \"kubernetes.io/projected/43d37775-34fa-4b0b-b56a-786ef557d7a9-kube-api-access-hjqk7\") pod \"43d37775-34fa-4b0b-b56a-786ef557d7a9\" (UID: \"43d37775-34fa-4b0b-b56a-786ef557d7a9\") " Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.566987 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43d37775-34fa-4b0b-b56a-786ef557d7a9-utilities" (OuterVolumeSpecName: "utilities") pod "43d37775-34fa-4b0b-b56a-786ef557d7a9" (UID: "43d37775-34fa-4b0b-b56a-786ef557d7a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.572062 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43d37775-34fa-4b0b-b56a-786ef557d7a9-kube-api-access-hjqk7" (OuterVolumeSpecName: "kube-api-access-hjqk7") pod "43d37775-34fa-4b0b-b56a-786ef557d7a9" (UID: "43d37775-34fa-4b0b-b56a-786ef557d7a9"). InnerVolumeSpecName "kube-api-access-hjqk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.669764 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43d37775-34fa-4b0b-b56a-786ef557d7a9-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.669822 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjqk7\" (UniqueName: \"kubernetes.io/projected/43d37775-34fa-4b0b-b56a-786ef557d7a9-kube-api-access-hjqk7\") on node \"crc\" DevicePath \"\"" Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.687063 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43d37775-34fa-4b0b-b56a-786ef557d7a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "43d37775-34fa-4b0b-b56a-786ef557d7a9" (UID: "43d37775-34fa-4b0b-b56a-786ef557d7a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.772373 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43d37775-34fa-4b0b-b56a-786ef557d7a9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.823392 4824 generic.go:334] "Generic (PLEG): container finished" podID="43d37775-34fa-4b0b-b56a-786ef557d7a9" containerID="5c7b1732a6d57c9bdc63b0f4d01d80f3bdc7dbc43fb7217b252ae5243e594470" exitCode=0 Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.823440 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmvc6" event={"ID":"43d37775-34fa-4b0b-b56a-786ef557d7a9","Type":"ContainerDied","Data":"5c7b1732a6d57c9bdc63b0f4d01d80f3bdc7dbc43fb7217b252ae5243e594470"} Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.823447 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dmvc6" Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.823479 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmvc6" event={"ID":"43d37775-34fa-4b0b-b56a-786ef557d7a9","Type":"ContainerDied","Data":"7ba88db7a96c344a847f5509fd56bc6dc45203a2802708be6f72a671d2d383ff"} Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.823497 4824 scope.go:117] "RemoveContainer" containerID="5c7b1732a6d57c9bdc63b0f4d01d80f3bdc7dbc43fb7217b252ae5243e594470" Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.848251 4824 scope.go:117] "RemoveContainer" containerID="5e2aa93540e7a055512dc5e7b7b79eedd0aeb9c3f72004a9c28f848e059735d2" Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.876214 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dmvc6"] Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.895552 4824 scope.go:117] "RemoveContainer" containerID="23ae8f7deb3c111901743ab8b96b40f0f6d61167f8f406b24b13ce3fb3471697" Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.896258 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dmvc6"] Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.925577 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43d37775-34fa-4b0b-b56a-786ef557d7a9" path="/var/lib/kubelet/pods/43d37775-34fa-4b0b-b56a-786ef557d7a9/volumes" Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.940563 4824 scope.go:117] "RemoveContainer" containerID="5c7b1732a6d57c9bdc63b0f4d01d80f3bdc7dbc43fb7217b252ae5243e594470" Dec 09 10:22:25 crc kubenswrapper[4824]: E1209 10:22:25.941354 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c7b1732a6d57c9bdc63b0f4d01d80f3bdc7dbc43fb7217b252ae5243e594470\": container with ID starting with 5c7b1732a6d57c9bdc63b0f4d01d80f3bdc7dbc43fb7217b252ae5243e594470 not found: ID does not exist" containerID="5c7b1732a6d57c9bdc63b0f4d01d80f3bdc7dbc43fb7217b252ae5243e594470" Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.941416 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c7b1732a6d57c9bdc63b0f4d01d80f3bdc7dbc43fb7217b252ae5243e594470"} err="failed to get container status \"5c7b1732a6d57c9bdc63b0f4d01d80f3bdc7dbc43fb7217b252ae5243e594470\": rpc error: code = NotFound desc = could not find container \"5c7b1732a6d57c9bdc63b0f4d01d80f3bdc7dbc43fb7217b252ae5243e594470\": container with ID starting with 5c7b1732a6d57c9bdc63b0f4d01d80f3bdc7dbc43fb7217b252ae5243e594470 not found: ID does not exist" Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.941457 4824 scope.go:117] "RemoveContainer" containerID="5e2aa93540e7a055512dc5e7b7b79eedd0aeb9c3f72004a9c28f848e059735d2" Dec 09 10:22:25 crc kubenswrapper[4824]: E1209 10:22:25.942062 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e2aa93540e7a055512dc5e7b7b79eedd0aeb9c3f72004a9c28f848e059735d2\": container with ID starting with 5e2aa93540e7a055512dc5e7b7b79eedd0aeb9c3f72004a9c28f848e059735d2 not found: ID does not exist" containerID="5e2aa93540e7a055512dc5e7b7b79eedd0aeb9c3f72004a9c28f848e059735d2" Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.942121 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e2aa93540e7a055512dc5e7b7b79eedd0aeb9c3f72004a9c28f848e059735d2"} err="failed to get container status \"5e2aa93540e7a055512dc5e7b7b79eedd0aeb9c3f72004a9c28f848e059735d2\": rpc error: code = NotFound desc = could not find container \"5e2aa93540e7a055512dc5e7b7b79eedd0aeb9c3f72004a9c28f848e059735d2\": container with ID starting with 5e2aa93540e7a055512dc5e7b7b79eedd0aeb9c3f72004a9c28f848e059735d2 not found: ID does not exist" Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.942157 4824 scope.go:117] "RemoveContainer" containerID="23ae8f7deb3c111901743ab8b96b40f0f6d61167f8f406b24b13ce3fb3471697" Dec 09 10:22:25 crc kubenswrapper[4824]: E1209 10:22:25.942702 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23ae8f7deb3c111901743ab8b96b40f0f6d61167f8f406b24b13ce3fb3471697\": container with ID starting with 23ae8f7deb3c111901743ab8b96b40f0f6d61167f8f406b24b13ce3fb3471697 not found: ID does not exist" containerID="23ae8f7deb3c111901743ab8b96b40f0f6d61167f8f406b24b13ce3fb3471697" Dec 09 10:22:25 crc kubenswrapper[4824]: I1209 10:22:25.942834 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23ae8f7deb3c111901743ab8b96b40f0f6d61167f8f406b24b13ce3fb3471697"} err="failed to get container status \"23ae8f7deb3c111901743ab8b96b40f0f6d61167f8f406b24b13ce3fb3471697\": rpc error: code = NotFound desc = could not find container \"23ae8f7deb3c111901743ab8b96b40f0f6d61167f8f406b24b13ce3fb3471697\": container with ID starting with 23ae8f7deb3c111901743ab8b96b40f0f6d61167f8f406b24b13ce3fb3471697 not found: ID does not exist" Dec 09 10:22:27 crc kubenswrapper[4824]: I1209 10:22:27.056433 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-jghjq"] Dec 09 10:22:27 crc kubenswrapper[4824]: I1209 10:22:27.076875 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-jghjq"] Dec 09 10:22:27 crc kubenswrapper[4824]: I1209 10:22:27.925095 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbceafb7-272a-45fc-9b56-df631986f90a" path="/var/lib/kubelet/pods/dbceafb7-272a-45fc-9b56-df631986f90a/volumes" Dec 09 10:23:02 crc kubenswrapper[4824]: I1209 10:23:02.861056 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:23:02 crc kubenswrapper[4824]: I1209 10:23:02.863014 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:23:09 crc kubenswrapper[4824]: I1209 10:23:09.671951 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-frnh4"] Dec 09 10:23:09 crc kubenswrapper[4824]: E1209 10:23:09.673283 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43d37775-34fa-4b0b-b56a-786ef557d7a9" containerName="extract-content" Dec 09 10:23:09 crc kubenswrapper[4824]: I1209 10:23:09.673306 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="43d37775-34fa-4b0b-b56a-786ef557d7a9" containerName="extract-content" Dec 09 10:23:09 crc kubenswrapper[4824]: E1209 10:23:09.673347 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43d37775-34fa-4b0b-b56a-786ef557d7a9" containerName="extract-utilities" Dec 09 10:23:09 crc kubenswrapper[4824]: I1209 10:23:09.673387 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="43d37775-34fa-4b0b-b56a-786ef557d7a9" containerName="extract-utilities" Dec 09 10:23:09 crc kubenswrapper[4824]: E1209 10:23:09.673406 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43d37775-34fa-4b0b-b56a-786ef557d7a9" containerName="registry-server" Dec 09 10:23:09 crc kubenswrapper[4824]: I1209 10:23:09.673416 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="43d37775-34fa-4b0b-b56a-786ef557d7a9" containerName="registry-server" Dec 09 10:23:09 crc kubenswrapper[4824]: I1209 10:23:09.673700 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="43d37775-34fa-4b0b-b56a-786ef557d7a9" containerName="registry-server" Dec 09 10:23:09 crc kubenswrapper[4824]: I1209 10:23:09.675675 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-frnh4" Dec 09 10:23:09 crc kubenswrapper[4824]: I1209 10:23:09.697292 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-frnh4"] Dec 09 10:23:09 crc kubenswrapper[4824]: I1209 10:23:09.731425 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x62r\" (UniqueName: \"kubernetes.io/projected/2205c20a-945a-49e3-85d3-dc8e9750c635-kube-api-access-4x62r\") pod \"community-operators-frnh4\" (UID: \"2205c20a-945a-49e3-85d3-dc8e9750c635\") " pod="openshift-marketplace/community-operators-frnh4" Dec 09 10:23:09 crc kubenswrapper[4824]: I1209 10:23:09.731910 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2205c20a-945a-49e3-85d3-dc8e9750c635-catalog-content\") pod \"community-operators-frnh4\" (UID: \"2205c20a-945a-49e3-85d3-dc8e9750c635\") " pod="openshift-marketplace/community-operators-frnh4" Dec 09 10:23:09 crc kubenswrapper[4824]: I1209 10:23:09.731962 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2205c20a-945a-49e3-85d3-dc8e9750c635-utilities\") pod \"community-operators-frnh4\" (UID: \"2205c20a-945a-49e3-85d3-dc8e9750c635\") " pod="openshift-marketplace/community-operators-frnh4" Dec 09 10:23:09 crc kubenswrapper[4824]: I1209 10:23:09.835422 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x62r\" (UniqueName: \"kubernetes.io/projected/2205c20a-945a-49e3-85d3-dc8e9750c635-kube-api-access-4x62r\") pod \"community-operators-frnh4\" (UID: \"2205c20a-945a-49e3-85d3-dc8e9750c635\") " pod="openshift-marketplace/community-operators-frnh4" Dec 09 10:23:09 crc kubenswrapper[4824]: I1209 10:23:09.835598 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2205c20a-945a-49e3-85d3-dc8e9750c635-catalog-content\") pod \"community-operators-frnh4\" (UID: \"2205c20a-945a-49e3-85d3-dc8e9750c635\") " pod="openshift-marketplace/community-operators-frnh4" Dec 09 10:23:09 crc kubenswrapper[4824]: I1209 10:23:09.835636 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2205c20a-945a-49e3-85d3-dc8e9750c635-utilities\") pod \"community-operators-frnh4\" (UID: \"2205c20a-945a-49e3-85d3-dc8e9750c635\") " pod="openshift-marketplace/community-operators-frnh4" Dec 09 10:23:09 crc kubenswrapper[4824]: I1209 10:23:09.836134 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2205c20a-945a-49e3-85d3-dc8e9750c635-catalog-content\") pod \"community-operators-frnh4\" (UID: \"2205c20a-945a-49e3-85d3-dc8e9750c635\") " pod="openshift-marketplace/community-operators-frnh4" Dec 09 10:23:09 crc kubenswrapper[4824]: I1209 10:23:09.836278 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2205c20a-945a-49e3-85d3-dc8e9750c635-utilities\") pod \"community-operators-frnh4\" (UID: \"2205c20a-945a-49e3-85d3-dc8e9750c635\") " pod="openshift-marketplace/community-operators-frnh4" Dec 09 10:23:09 crc kubenswrapper[4824]: I1209 10:23:09.866605 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x62r\" (UniqueName: \"kubernetes.io/projected/2205c20a-945a-49e3-85d3-dc8e9750c635-kube-api-access-4x62r\") pod \"community-operators-frnh4\" (UID: \"2205c20a-945a-49e3-85d3-dc8e9750c635\") " pod="openshift-marketplace/community-operators-frnh4" Dec 09 10:23:10 crc kubenswrapper[4824]: I1209 10:23:10.001535 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-frnh4" Dec 09 10:23:10 crc kubenswrapper[4824]: I1209 10:23:10.644932 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-frnh4"] Dec 09 10:23:11 crc kubenswrapper[4824]: I1209 10:23:11.561252 4824 generic.go:334] "Generic (PLEG): container finished" podID="2205c20a-945a-49e3-85d3-dc8e9750c635" containerID="c5a3a85570f4cb4dc871fbbe509ebf484b1f376aa51395f52bdfa18f540a98f3" exitCode=0 Dec 09 10:23:11 crc kubenswrapper[4824]: I1209 10:23:11.561295 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-frnh4" event={"ID":"2205c20a-945a-49e3-85d3-dc8e9750c635","Type":"ContainerDied","Data":"c5a3a85570f4cb4dc871fbbe509ebf484b1f376aa51395f52bdfa18f540a98f3"} Dec 09 10:23:11 crc kubenswrapper[4824]: I1209 10:23:11.561718 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-frnh4" event={"ID":"2205c20a-945a-49e3-85d3-dc8e9750c635","Type":"ContainerStarted","Data":"11e9e1a644731ad1d2659886eb41afd06248fb99fb5bd2d583ce50745261fc7d"} Dec 09 10:23:11 crc kubenswrapper[4824]: I1209 10:23:11.563607 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 10:23:12 crc kubenswrapper[4824]: I1209 10:23:12.575202 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-frnh4" event={"ID":"2205c20a-945a-49e3-85d3-dc8e9750c635","Type":"ContainerStarted","Data":"e21ba35e0256a75c9eb69444ab9f7f83b096fafc3f4e20e2c7353f0bfaa1cb22"} Dec 09 10:23:13 crc kubenswrapper[4824]: I1209 10:23:13.598440 4824 generic.go:334] "Generic (PLEG): container finished" podID="2205c20a-945a-49e3-85d3-dc8e9750c635" containerID="e21ba35e0256a75c9eb69444ab9f7f83b096fafc3f4e20e2c7353f0bfaa1cb22" exitCode=0 Dec 09 10:23:13 crc kubenswrapper[4824]: I1209 10:23:13.598740 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-frnh4" event={"ID":"2205c20a-945a-49e3-85d3-dc8e9750c635","Type":"ContainerDied","Data":"e21ba35e0256a75c9eb69444ab9f7f83b096fafc3f4e20e2c7353f0bfaa1cb22"} Dec 09 10:23:14 crc kubenswrapper[4824]: I1209 10:23:14.628246 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-frnh4" event={"ID":"2205c20a-945a-49e3-85d3-dc8e9750c635","Type":"ContainerStarted","Data":"ff9a28c1ee09a0bbb13f671a4ab8f2be32f36229aeaa7b4ed8a750613045669e"} Dec 09 10:23:14 crc kubenswrapper[4824]: I1209 10:23:14.661638 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-frnh4" podStartSLOduration=3.086148206 podStartE2EDuration="5.661611358s" podCreationTimestamp="2025-12-09 10:23:09 +0000 UTC" firstStartedPulling="2025-12-09 10:23:11.563367156 +0000 UTC m=+2147.897871813" lastFinishedPulling="2025-12-09 10:23:14.138830278 +0000 UTC m=+2150.473334965" observedRunningTime="2025-12-09 10:23:14.647516964 +0000 UTC m=+2150.982021641" watchObservedRunningTime="2025-12-09 10:23:14.661611358 +0000 UTC m=+2150.996116025" Dec 09 10:23:16 crc kubenswrapper[4824]: I1209 10:23:16.282603 4824 scope.go:117] "RemoveContainer" containerID="751410b43bd034373b18f9792b5d576ad6b6aa7d073a4a19569c94413f9c45c0" Dec 09 10:23:20 crc kubenswrapper[4824]: I1209 10:23:20.001688 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-frnh4" Dec 09 10:23:20 crc kubenswrapper[4824]: I1209 10:23:20.002028 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-frnh4" Dec 09 10:23:20 crc kubenswrapper[4824]: I1209 10:23:20.066632 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-frnh4" Dec 09 10:23:20 crc kubenswrapper[4824]: I1209 10:23:20.742817 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-frnh4" Dec 09 10:23:20 crc kubenswrapper[4824]: I1209 10:23:20.798454 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-frnh4"] Dec 09 10:23:22 crc kubenswrapper[4824]: I1209 10:23:22.711952 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-frnh4" podUID="2205c20a-945a-49e3-85d3-dc8e9750c635" containerName="registry-server" containerID="cri-o://ff9a28c1ee09a0bbb13f671a4ab8f2be32f36229aeaa7b4ed8a750613045669e" gracePeriod=2 Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.291930 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-frnh4" Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.426336 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2205c20a-945a-49e3-85d3-dc8e9750c635-catalog-content\") pod \"2205c20a-945a-49e3-85d3-dc8e9750c635\" (UID: \"2205c20a-945a-49e3-85d3-dc8e9750c635\") " Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.426393 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4x62r\" (UniqueName: \"kubernetes.io/projected/2205c20a-945a-49e3-85d3-dc8e9750c635-kube-api-access-4x62r\") pod \"2205c20a-945a-49e3-85d3-dc8e9750c635\" (UID: \"2205c20a-945a-49e3-85d3-dc8e9750c635\") " Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.427328 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2205c20a-945a-49e3-85d3-dc8e9750c635-utilities\") pod \"2205c20a-945a-49e3-85d3-dc8e9750c635\" (UID: \"2205c20a-945a-49e3-85d3-dc8e9750c635\") " Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.427473 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2205c20a-945a-49e3-85d3-dc8e9750c635-utilities" (OuterVolumeSpecName: "utilities") pod "2205c20a-945a-49e3-85d3-dc8e9750c635" (UID: "2205c20a-945a-49e3-85d3-dc8e9750c635"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.428113 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2205c20a-945a-49e3-85d3-dc8e9750c635-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.432285 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2205c20a-945a-49e3-85d3-dc8e9750c635-kube-api-access-4x62r" (OuterVolumeSpecName: "kube-api-access-4x62r") pod "2205c20a-945a-49e3-85d3-dc8e9750c635" (UID: "2205c20a-945a-49e3-85d3-dc8e9750c635"). InnerVolumeSpecName "kube-api-access-4x62r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.479580 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2205c20a-945a-49e3-85d3-dc8e9750c635-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2205c20a-945a-49e3-85d3-dc8e9750c635" (UID: "2205c20a-945a-49e3-85d3-dc8e9750c635"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.532052 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2205c20a-945a-49e3-85d3-dc8e9750c635-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.532091 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4x62r\" (UniqueName: \"kubernetes.io/projected/2205c20a-945a-49e3-85d3-dc8e9750c635-kube-api-access-4x62r\") on node \"crc\" DevicePath \"\"" Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.736297 4824 generic.go:334] "Generic (PLEG): container finished" podID="2205c20a-945a-49e3-85d3-dc8e9750c635" containerID="ff9a28c1ee09a0bbb13f671a4ab8f2be32f36229aeaa7b4ed8a750613045669e" exitCode=0 Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.736363 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-frnh4" event={"ID":"2205c20a-945a-49e3-85d3-dc8e9750c635","Type":"ContainerDied","Data":"ff9a28c1ee09a0bbb13f671a4ab8f2be32f36229aeaa7b4ed8a750613045669e"} Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.736441 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-frnh4" event={"ID":"2205c20a-945a-49e3-85d3-dc8e9750c635","Type":"ContainerDied","Data":"11e9e1a644731ad1d2659886eb41afd06248fb99fb5bd2d583ce50745261fc7d"} Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.736457 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-frnh4" Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.736494 4824 scope.go:117] "RemoveContainer" containerID="ff9a28c1ee09a0bbb13f671a4ab8f2be32f36229aeaa7b4ed8a750613045669e" Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.786536 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-frnh4"] Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.796115 4824 scope.go:117] "RemoveContainer" containerID="e21ba35e0256a75c9eb69444ab9f7f83b096fafc3f4e20e2c7353f0bfaa1cb22" Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.800136 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-frnh4"] Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.824864 4824 scope.go:117] "RemoveContainer" containerID="c5a3a85570f4cb4dc871fbbe509ebf484b1f376aa51395f52bdfa18f540a98f3" Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.891777 4824 scope.go:117] "RemoveContainer" containerID="ff9a28c1ee09a0bbb13f671a4ab8f2be32f36229aeaa7b4ed8a750613045669e" Dec 09 10:23:23 crc kubenswrapper[4824]: E1209 10:23:23.892385 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff9a28c1ee09a0bbb13f671a4ab8f2be32f36229aeaa7b4ed8a750613045669e\": container with ID starting with ff9a28c1ee09a0bbb13f671a4ab8f2be32f36229aeaa7b4ed8a750613045669e not found: ID does not exist" containerID="ff9a28c1ee09a0bbb13f671a4ab8f2be32f36229aeaa7b4ed8a750613045669e" Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.892437 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff9a28c1ee09a0bbb13f671a4ab8f2be32f36229aeaa7b4ed8a750613045669e"} err="failed to get container status \"ff9a28c1ee09a0bbb13f671a4ab8f2be32f36229aeaa7b4ed8a750613045669e\": rpc error: code = NotFound desc = could not find container \"ff9a28c1ee09a0bbb13f671a4ab8f2be32f36229aeaa7b4ed8a750613045669e\": container with ID starting with ff9a28c1ee09a0bbb13f671a4ab8f2be32f36229aeaa7b4ed8a750613045669e not found: ID does not exist" Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.892471 4824 scope.go:117] "RemoveContainer" containerID="e21ba35e0256a75c9eb69444ab9f7f83b096fafc3f4e20e2c7353f0bfaa1cb22" Dec 09 10:23:23 crc kubenswrapper[4824]: E1209 10:23:23.892844 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e21ba35e0256a75c9eb69444ab9f7f83b096fafc3f4e20e2c7353f0bfaa1cb22\": container with ID starting with e21ba35e0256a75c9eb69444ab9f7f83b096fafc3f4e20e2c7353f0bfaa1cb22 not found: ID does not exist" containerID="e21ba35e0256a75c9eb69444ab9f7f83b096fafc3f4e20e2c7353f0bfaa1cb22" Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.893938 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e21ba35e0256a75c9eb69444ab9f7f83b096fafc3f4e20e2c7353f0bfaa1cb22"} err="failed to get container status \"e21ba35e0256a75c9eb69444ab9f7f83b096fafc3f4e20e2c7353f0bfaa1cb22\": rpc error: code = NotFound desc = could not find container \"e21ba35e0256a75c9eb69444ab9f7f83b096fafc3f4e20e2c7353f0bfaa1cb22\": container with ID starting with e21ba35e0256a75c9eb69444ab9f7f83b096fafc3f4e20e2c7353f0bfaa1cb22 not found: ID does not exist" Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.893962 4824 scope.go:117] "RemoveContainer" containerID="c5a3a85570f4cb4dc871fbbe509ebf484b1f376aa51395f52bdfa18f540a98f3" Dec 09 10:23:23 crc kubenswrapper[4824]: E1209 10:23:23.895051 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5a3a85570f4cb4dc871fbbe509ebf484b1f376aa51395f52bdfa18f540a98f3\": container with ID starting with c5a3a85570f4cb4dc871fbbe509ebf484b1f376aa51395f52bdfa18f540a98f3 not found: ID does not exist" containerID="c5a3a85570f4cb4dc871fbbe509ebf484b1f376aa51395f52bdfa18f540a98f3" Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.895081 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5a3a85570f4cb4dc871fbbe509ebf484b1f376aa51395f52bdfa18f540a98f3"} err="failed to get container status \"c5a3a85570f4cb4dc871fbbe509ebf484b1f376aa51395f52bdfa18f540a98f3\": rpc error: code = NotFound desc = could not find container \"c5a3a85570f4cb4dc871fbbe509ebf484b1f376aa51395f52bdfa18f540a98f3\": container with ID starting with c5a3a85570f4cb4dc871fbbe509ebf484b1f376aa51395f52bdfa18f540a98f3 not found: ID does not exist" Dec 09 10:23:23 crc kubenswrapper[4824]: I1209 10:23:23.928144 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2205c20a-945a-49e3-85d3-dc8e9750c635" path="/var/lib/kubelet/pods/2205c20a-945a-49e3-85d3-dc8e9750c635/volumes" Dec 09 10:23:32 crc kubenswrapper[4824]: I1209 10:23:32.860860 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:23:32 crc kubenswrapper[4824]: I1209 10:23:32.861389 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:23:35 crc kubenswrapper[4824]: I1209 10:23:35.121564 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-366d-account-create-update-zfvhr"] Dec 09 10:23:35 crc kubenswrapper[4824]: I1209 10:23:35.137008 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-366d-account-create-update-zfvhr"] Dec 09 10:23:35 crc kubenswrapper[4824]: I1209 10:23:35.942834 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e79ef98a-aa4c-4452-b217-bbb43c73fed3" path="/var/lib/kubelet/pods/e79ef98a-aa4c-4452-b217-bbb43c73fed3/volumes" Dec 09 10:23:36 crc kubenswrapper[4824]: I1209 10:23:36.042966 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-bzx8k"] Dec 09 10:23:36 crc kubenswrapper[4824]: I1209 10:23:36.062464 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-2966b"] Dec 09 10:23:36 crc kubenswrapper[4824]: I1209 10:23:36.073323 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-bzx8k"] Dec 09 10:23:36 crc kubenswrapper[4824]: I1209 10:23:36.085812 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-7461-account-create-update-gvr6w"] Dec 09 10:23:36 crc kubenswrapper[4824]: I1209 10:23:36.097563 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-840e-account-create-update-9f4xh"] Dec 09 10:23:36 crc kubenswrapper[4824]: I1209 10:23:36.110637 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-h9wzl"] Dec 09 10:23:36 crc kubenswrapper[4824]: I1209 10:23:36.122485 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-2966b"] Dec 09 10:23:36 crc kubenswrapper[4824]: I1209 10:23:36.133317 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-h9wzl"] Dec 09 10:23:36 crc kubenswrapper[4824]: I1209 10:23:36.145473 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-840e-account-create-update-9f4xh"] Dec 09 10:23:36 crc kubenswrapper[4824]: I1209 10:23:36.156545 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-7461-account-create-update-gvr6w"] Dec 09 10:23:37 crc kubenswrapper[4824]: I1209 10:23:37.923766 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52e9f723-e235-4411-aeff-818c586afe9d" path="/var/lib/kubelet/pods/52e9f723-e235-4411-aeff-818c586afe9d/volumes" Dec 09 10:23:37 crc kubenswrapper[4824]: I1209 10:23:37.925047 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9658659d-f0c7-47b9-95ef-14a8f4be023e" path="/var/lib/kubelet/pods/9658659d-f0c7-47b9-95ef-14a8f4be023e/volumes" Dec 09 10:23:37 crc kubenswrapper[4824]: I1209 10:23:37.925726 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4668f61-ef52-4072-bd2c-ed91c7298ed3" path="/var/lib/kubelet/pods/a4668f61-ef52-4072-bd2c-ed91c7298ed3/volumes" Dec 09 10:23:37 crc kubenswrapper[4824]: I1209 10:23:37.926446 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee" path="/var/lib/kubelet/pods/ae21e48f-0fe9-4ce0-aaf2-aa2c3119fbee/volumes" Dec 09 10:23:37 crc kubenswrapper[4824]: I1209 10:23:37.927591 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cff7cbdd-8ed6-48e5-9fb0-35123690fac0" path="/var/lib/kubelet/pods/cff7cbdd-8ed6-48e5-9fb0-35123690fac0/volumes" Dec 09 10:24:02 crc kubenswrapper[4824]: I1209 10:24:02.860852 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:24:02 crc kubenswrapper[4824]: I1209 10:24:02.861590 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:24:02 crc kubenswrapper[4824]: I1209 10:24:02.861660 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 10:24:02 crc kubenswrapper[4824]: I1209 10:24:02.863323 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d4d2394221cb52bd0aff000b3ed8a1bb39ce26a5dee3baa44cc5e9dfefc07899"} pod="openshift-machine-config-operator/machine-config-daemon-dth8x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 10:24:02 crc kubenswrapper[4824]: I1209 10:24:02.863417 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" containerID="cri-o://d4d2394221cb52bd0aff000b3ed8a1bb39ce26a5dee3baa44cc5e9dfefc07899" gracePeriod=600 Dec 09 10:24:03 crc kubenswrapper[4824]: I1209 10:24:03.166435 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerID="d4d2394221cb52bd0aff000b3ed8a1bb39ce26a5dee3baa44cc5e9dfefc07899" exitCode=0 Dec 09 10:24:03 crc kubenswrapper[4824]: I1209 10:24:03.166484 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerDied","Data":"d4d2394221cb52bd0aff000b3ed8a1bb39ce26a5dee3baa44cc5e9dfefc07899"} Dec 09 10:24:03 crc kubenswrapper[4824]: I1209 10:24:03.166536 4824 scope.go:117] "RemoveContainer" containerID="7a78531c9e1d440d986c712476febe2cecc2b3e3d8841b518afaffe2b0231cb1" Dec 09 10:24:04 crc kubenswrapper[4824]: I1209 10:24:04.180658 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerStarted","Data":"2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008"} Dec 09 10:24:07 crc kubenswrapper[4824]: I1209 10:24:07.215495 4824 generic.go:334] "Generic (PLEG): container finished" podID="4a2f92dc-7710-4e22-8e1a-d303580314c2" containerID="bcba845388f43a065c6411df438419dc8a6000ec29e4cf23a7f22bc329a63294" exitCode=0 Dec 09 10:24:07 crc kubenswrapper[4824]: I1209 10:24:07.215584 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs" event={"ID":"4a2f92dc-7710-4e22-8e1a-d303580314c2","Type":"ContainerDied","Data":"bcba845388f43a065c6411df438419dc8a6000ec29e4cf23a7f22bc329a63294"} Dec 09 10:24:08 crc kubenswrapper[4824]: I1209 10:24:08.747512 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs" Dec 09 10:24:08 crc kubenswrapper[4824]: I1209 10:24:08.861674 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwlvr\" (UniqueName: \"kubernetes.io/projected/4a2f92dc-7710-4e22-8e1a-d303580314c2-kube-api-access-vwlvr\") pod \"4a2f92dc-7710-4e22-8e1a-d303580314c2\" (UID: \"4a2f92dc-7710-4e22-8e1a-d303580314c2\") " Dec 09 10:24:08 crc kubenswrapper[4824]: I1209 10:24:08.861735 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a2f92dc-7710-4e22-8e1a-d303580314c2-inventory\") pod \"4a2f92dc-7710-4e22-8e1a-d303580314c2\" (UID: \"4a2f92dc-7710-4e22-8e1a-d303580314c2\") " Dec 09 10:24:08 crc kubenswrapper[4824]: I1209 10:24:08.861921 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a2f92dc-7710-4e22-8e1a-d303580314c2-ssh-key\") pod \"4a2f92dc-7710-4e22-8e1a-d303580314c2\" (UID: \"4a2f92dc-7710-4e22-8e1a-d303580314c2\") " Dec 09 10:24:08 crc kubenswrapper[4824]: I1209 10:24:08.867507 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a2f92dc-7710-4e22-8e1a-d303580314c2-kube-api-access-vwlvr" (OuterVolumeSpecName: "kube-api-access-vwlvr") pod "4a2f92dc-7710-4e22-8e1a-d303580314c2" (UID: "4a2f92dc-7710-4e22-8e1a-d303580314c2"). InnerVolumeSpecName "kube-api-access-vwlvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:24:08 crc kubenswrapper[4824]: I1209 10:24:08.899342 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a2f92dc-7710-4e22-8e1a-d303580314c2-inventory" (OuterVolumeSpecName: "inventory") pod "4a2f92dc-7710-4e22-8e1a-d303580314c2" (UID: "4a2f92dc-7710-4e22-8e1a-d303580314c2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:24:08 crc kubenswrapper[4824]: I1209 10:24:08.899414 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a2f92dc-7710-4e22-8e1a-d303580314c2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4a2f92dc-7710-4e22-8e1a-d303580314c2" (UID: "4a2f92dc-7710-4e22-8e1a-d303580314c2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:24:08 crc kubenswrapper[4824]: I1209 10:24:08.964820 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwlvr\" (UniqueName: \"kubernetes.io/projected/4a2f92dc-7710-4e22-8e1a-d303580314c2-kube-api-access-vwlvr\") on node \"crc\" DevicePath \"\"" Dec 09 10:24:08 crc kubenswrapper[4824]: I1209 10:24:08.964857 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a2f92dc-7710-4e22-8e1a-d303580314c2-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 10:24:08 crc kubenswrapper[4824]: I1209 10:24:08.964865 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a2f92dc-7710-4e22-8e1a-d303580314c2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.239960 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs" event={"ID":"4a2f92dc-7710-4e22-8e1a-d303580314c2","Type":"ContainerDied","Data":"69f927d52786e74a048727d53815a0605b18cac2e07e01265d1d50ef62ca1ac7"} Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.240001 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69f927d52786e74a048727d53815a0605b18cac2e07e01265d1d50ef62ca1ac7" Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.240034 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs" Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.340391 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq"] Dec 09 10:24:09 crc kubenswrapper[4824]: E1209 10:24:09.341413 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2205c20a-945a-49e3-85d3-dc8e9750c635" containerName="registry-server" Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.341456 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2205c20a-945a-49e3-85d3-dc8e9750c635" containerName="registry-server" Dec 09 10:24:09 crc kubenswrapper[4824]: E1209 10:24:09.341500 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2205c20a-945a-49e3-85d3-dc8e9750c635" containerName="extract-content" Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.341509 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2205c20a-945a-49e3-85d3-dc8e9750c635" containerName="extract-content" Dec 09 10:24:09 crc kubenswrapper[4824]: E1209 10:24:09.341557 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2205c20a-945a-49e3-85d3-dc8e9750c635" containerName="extract-utilities" Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.341566 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2205c20a-945a-49e3-85d3-dc8e9750c635" containerName="extract-utilities" Dec 09 10:24:09 crc kubenswrapper[4824]: E1209 10:24:09.341582 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a2f92dc-7710-4e22-8e1a-d303580314c2" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.341591 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a2f92dc-7710-4e22-8e1a-d303580314c2" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.341947 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2205c20a-945a-49e3-85d3-dc8e9750c635" containerName="registry-server" Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.341974 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a2f92dc-7710-4e22-8e1a-d303580314c2" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.343167 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq" Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.346619 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.348218 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.348364 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5nzld" Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.353237 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq"] Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.365330 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.478406 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzxk8\" (UniqueName: \"kubernetes.io/projected/329f5680-bbd1-427c-b365-70c2109c214f-kube-api-access-bzxk8\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq\" (UID: \"329f5680-bbd1-427c-b365-70c2109c214f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq" Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.478477 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/329f5680-bbd1-427c-b365-70c2109c214f-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq\" (UID: \"329f5680-bbd1-427c-b365-70c2109c214f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq" Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.478687 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/329f5680-bbd1-427c-b365-70c2109c214f-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq\" (UID: \"329f5680-bbd1-427c-b365-70c2109c214f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq" Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.581385 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzxk8\" (UniqueName: \"kubernetes.io/projected/329f5680-bbd1-427c-b365-70c2109c214f-kube-api-access-bzxk8\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq\" (UID: \"329f5680-bbd1-427c-b365-70c2109c214f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq" Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.581735 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/329f5680-bbd1-427c-b365-70c2109c214f-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq\" (UID: \"329f5680-bbd1-427c-b365-70c2109c214f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq" Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.581917 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/329f5680-bbd1-427c-b365-70c2109c214f-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq\" (UID: \"329f5680-bbd1-427c-b365-70c2109c214f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq" Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.587174 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/329f5680-bbd1-427c-b365-70c2109c214f-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq\" (UID: \"329f5680-bbd1-427c-b365-70c2109c214f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq" Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.587215 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/329f5680-bbd1-427c-b365-70c2109c214f-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq\" (UID: \"329f5680-bbd1-427c-b365-70c2109c214f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq" Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.599505 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzxk8\" (UniqueName: \"kubernetes.io/projected/329f5680-bbd1-427c-b365-70c2109c214f-kube-api-access-bzxk8\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq\" (UID: \"329f5680-bbd1-427c-b365-70c2109c214f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq" Dec 09 10:24:09 crc kubenswrapper[4824]: I1209 10:24:09.662219 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq" Dec 09 10:24:10 crc kubenswrapper[4824]: I1209 10:24:10.289154 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq"] Dec 09 10:24:11 crc kubenswrapper[4824]: I1209 10:24:11.262814 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq" event={"ID":"329f5680-bbd1-427c-b365-70c2109c214f","Type":"ContainerStarted","Data":"d82ae3190b453fb1751e4b66be787818c12dde50675218fd1a139cb7e009219e"} Dec 09 10:24:11 crc kubenswrapper[4824]: I1209 10:24:11.263355 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq" event={"ID":"329f5680-bbd1-427c-b365-70c2109c214f","Type":"ContainerStarted","Data":"03240501ddf1c53bb109f73bc4a8ba39dc026ac9b5df77e53f75491065f3049d"} Dec 09 10:24:11 crc kubenswrapper[4824]: I1209 10:24:11.295667 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq" podStartSLOduration=1.808511587 podStartE2EDuration="2.295638044s" podCreationTimestamp="2025-12-09 10:24:09 +0000 UTC" firstStartedPulling="2025-12-09 10:24:10.324981684 +0000 UTC m=+2206.659486351" lastFinishedPulling="2025-12-09 10:24:10.812108121 +0000 UTC m=+2207.146612808" observedRunningTime="2025-12-09 10:24:11.280215448 +0000 UTC m=+2207.614720125" watchObservedRunningTime="2025-12-09 10:24:11.295638044 +0000 UTC m=+2207.630142711" Dec 09 10:24:12 crc kubenswrapper[4824]: I1209 10:24:12.046133 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8q4b9"] Dec 09 10:24:12 crc kubenswrapper[4824]: I1209 10:24:12.057014 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8q4b9"] Dec 09 10:24:13 crc kubenswrapper[4824]: I1209 10:24:13.926705 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="221367da-a9dd-488d-a2c5-69bc0f57e589" path="/var/lib/kubelet/pods/221367da-a9dd-488d-a2c5-69bc0f57e589/volumes" Dec 09 10:24:16 crc kubenswrapper[4824]: I1209 10:24:16.373625 4824 scope.go:117] "RemoveContainer" containerID="bc239bc9d807eae04960e5f9ed3cb500ed60adef14d484a2d64e327fee158164" Dec 09 10:24:16 crc kubenswrapper[4824]: I1209 10:24:16.423561 4824 scope.go:117] "RemoveContainer" containerID="525064878d70a2f6e2ffadc3478f6b6c99c7a46e197c0d269b691fc42ec86995" Dec 09 10:24:16 crc kubenswrapper[4824]: I1209 10:24:16.481346 4824 scope.go:117] "RemoveContainer" containerID="70b7050442fdf6eb10381f437e45fa00761a5e37ce90432c54bcefc8cd9db7b2" Dec 09 10:24:16 crc kubenswrapper[4824]: I1209 10:24:16.551806 4824 scope.go:117] "RemoveContainer" containerID="a6e8c0b9da4a4795c1e6d82834746a67795b93a047b9b70e2cc2c184147b7013" Dec 09 10:24:16 crc kubenswrapper[4824]: I1209 10:24:16.624641 4824 scope.go:117] "RemoveContainer" containerID="b252953e0c9dd129c85c067821d00db94fe90ad5cad332b4f2c3fb877c7d20bb" Dec 09 10:24:16 crc kubenswrapper[4824]: I1209 10:24:16.680544 4824 scope.go:117] "RemoveContainer" containerID="1f146c01309ed0552d65c2763666643b8d29faf00a88737fd4ba628b79c8cd9e" Dec 09 10:24:16 crc kubenswrapper[4824]: I1209 10:24:16.737262 4824 scope.go:117] "RemoveContainer" containerID="77650c36dee46cfe0e7fdbcbd89b23ea80a5e7ae6ea5ba7c73f25de9c50b2f2c" Dec 09 10:24:28 crc kubenswrapper[4824]: I1209 10:24:28.065456 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-52bnj"] Dec 09 10:24:28 crc kubenswrapper[4824]: I1209 10:24:28.076700 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-52bnj"] Dec 09 10:24:29 crc kubenswrapper[4824]: I1209 10:24:29.926247 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3727faad-272d-4a6a-a2ba-7b5be05cd11b" path="/var/lib/kubelet/pods/3727faad-272d-4a6a-a2ba-7b5be05cd11b/volumes" Dec 09 10:24:31 crc kubenswrapper[4824]: I1209 10:24:31.037610 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-f925-account-create-update-zk59p"] Dec 09 10:24:31 crc kubenswrapper[4824]: I1209 10:24:31.050164 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-f925-account-create-update-zk59p"] Dec 09 10:24:31 crc kubenswrapper[4824]: I1209 10:24:31.925669 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96ede01d-246b-47e6-ac4d-06c74436006c" path="/var/lib/kubelet/pods/96ede01d-246b-47e6-ac4d-06c74436006c/volumes" Dec 09 10:24:50 crc kubenswrapper[4824]: I1209 10:24:50.052676 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-d6jn9"] Dec 09 10:24:50 crc kubenswrapper[4824]: I1209 10:24:50.068907 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rcv9n"] Dec 09 10:24:50 crc kubenswrapper[4824]: I1209 10:24:50.078945 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-d6jn9"] Dec 09 10:24:50 crc kubenswrapper[4824]: I1209 10:24:50.088394 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rcv9n"] Dec 09 10:24:51 crc kubenswrapper[4824]: I1209 10:24:51.922998 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5f03343-b401-422a-bea2-7c97f7a99600" path="/var/lib/kubelet/pods/b5f03343-b401-422a-bea2-7c97f7a99600/volumes" Dec 09 10:24:51 crc kubenswrapper[4824]: I1209 10:24:51.924064 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32" path="/var/lib/kubelet/pods/de1a0fbc-88eb-45b3-bd94-9ff44fbf7e32/volumes" Dec 09 10:25:16 crc kubenswrapper[4824]: I1209 10:25:16.934586 4824 scope.go:117] "RemoveContainer" containerID="36892ea000779b2d91103ba0acc4ac4bc431aef8ae2a7a1f5876cf8a9e8eee23" Dec 09 10:25:16 crc kubenswrapper[4824]: I1209 10:25:16.967237 4824 scope.go:117] "RemoveContainer" containerID="fe5fbf86df0368876aa815d6a76c6b9d94453af6aa6c53f5db8f896782593237" Dec 09 10:25:17 crc kubenswrapper[4824]: I1209 10:25:17.016461 4824 scope.go:117] "RemoveContainer" containerID="a263b2169fe43a5e644746c7440339c9006ef8de764e6dcbc5e1b876e301a6c7" Dec 09 10:25:17 crc kubenswrapper[4824]: I1209 10:25:17.079081 4824 scope.go:117] "RemoveContainer" containerID="2c46216c2b5f4579d9654864a4f17ec06cebbdf5bcccad4f0d54c87f26e021c3" Dec 09 10:25:33 crc kubenswrapper[4824]: I1209 10:25:33.058198 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-282jv"] Dec 09 10:25:33 crc kubenswrapper[4824]: I1209 10:25:33.091120 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-282jv"] Dec 09 10:25:33 crc kubenswrapper[4824]: I1209 10:25:33.926142 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a34cbc14-a8b4-4993-89ba-b36f8a9a4065" path="/var/lib/kubelet/pods/a34cbc14-a8b4-4993-89ba-b36f8a9a4065/volumes" Dec 09 10:25:40 crc kubenswrapper[4824]: I1209 10:25:40.318050 4824 generic.go:334] "Generic (PLEG): container finished" podID="329f5680-bbd1-427c-b365-70c2109c214f" containerID="d82ae3190b453fb1751e4b66be787818c12dde50675218fd1a139cb7e009219e" exitCode=0 Dec 09 10:25:40 crc kubenswrapper[4824]: I1209 10:25:40.318188 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq" event={"ID":"329f5680-bbd1-427c-b365-70c2109c214f","Type":"ContainerDied","Data":"d82ae3190b453fb1751e4b66be787818c12dde50675218fd1a139cb7e009219e"} Dec 09 10:25:41 crc kubenswrapper[4824]: I1209 10:25:41.806052 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq" Dec 09 10:25:41 crc kubenswrapper[4824]: I1209 10:25:41.903827 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/329f5680-bbd1-427c-b365-70c2109c214f-inventory\") pod \"329f5680-bbd1-427c-b365-70c2109c214f\" (UID: \"329f5680-bbd1-427c-b365-70c2109c214f\") " Dec 09 10:25:41 crc kubenswrapper[4824]: I1209 10:25:41.904603 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/329f5680-bbd1-427c-b365-70c2109c214f-ssh-key\") pod \"329f5680-bbd1-427c-b365-70c2109c214f\" (UID: \"329f5680-bbd1-427c-b365-70c2109c214f\") " Dec 09 10:25:41 crc kubenswrapper[4824]: I1209 10:25:41.904851 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzxk8\" (UniqueName: \"kubernetes.io/projected/329f5680-bbd1-427c-b365-70c2109c214f-kube-api-access-bzxk8\") pod \"329f5680-bbd1-427c-b365-70c2109c214f\" (UID: \"329f5680-bbd1-427c-b365-70c2109c214f\") " Dec 09 10:25:41 crc kubenswrapper[4824]: I1209 10:25:41.912988 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/329f5680-bbd1-427c-b365-70c2109c214f-kube-api-access-bzxk8" (OuterVolumeSpecName: "kube-api-access-bzxk8") pod "329f5680-bbd1-427c-b365-70c2109c214f" (UID: "329f5680-bbd1-427c-b365-70c2109c214f"). InnerVolumeSpecName "kube-api-access-bzxk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:25:41 crc kubenswrapper[4824]: I1209 10:25:41.940945 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/329f5680-bbd1-427c-b365-70c2109c214f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "329f5680-bbd1-427c-b365-70c2109c214f" (UID: "329f5680-bbd1-427c-b365-70c2109c214f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:25:41 crc kubenswrapper[4824]: I1209 10:25:41.943659 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/329f5680-bbd1-427c-b365-70c2109c214f-inventory" (OuterVolumeSpecName: "inventory") pod "329f5680-bbd1-427c-b365-70c2109c214f" (UID: "329f5680-bbd1-427c-b365-70c2109c214f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.008615 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/329f5680-bbd1-427c-b365-70c2109c214f-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.008895 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/329f5680-bbd1-427c-b365-70c2109c214f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.009018 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzxk8\" (UniqueName: \"kubernetes.io/projected/329f5680-bbd1-427c-b365-70c2109c214f-kube-api-access-bzxk8\") on node \"crc\" DevicePath \"\"" Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.351135 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq" event={"ID":"329f5680-bbd1-427c-b365-70c2109c214f","Type":"ContainerDied","Data":"03240501ddf1c53bb109f73bc4a8ba39dc026ac9b5df77e53f75491065f3049d"} Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.351220 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03240501ddf1c53bb109f73bc4a8ba39dc026ac9b5df77e53f75491065f3049d" Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.351324 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq" Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.466220 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-stmd9"] Dec 09 10:25:42 crc kubenswrapper[4824]: E1209 10:25:42.467247 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="329f5680-bbd1-427c-b365-70c2109c214f" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.467272 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="329f5680-bbd1-427c-b365-70c2109c214f" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.468461 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="329f5680-bbd1-427c-b365-70c2109c214f" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.469733 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-stmd9" Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.472932 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.473852 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.473879 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5nzld" Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.474022 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.493474 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-stmd9"] Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.542054 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dqtq\" (UniqueName: \"kubernetes.io/projected/00da1369-ca64-4df7-9924-fbcd75b38d86-kube-api-access-6dqtq\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-stmd9\" (UID: \"00da1369-ca64-4df7-9924-fbcd75b38d86\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-stmd9" Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.542576 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00da1369-ca64-4df7-9924-fbcd75b38d86-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-stmd9\" (UID: \"00da1369-ca64-4df7-9924-fbcd75b38d86\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-stmd9" Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.542728 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00da1369-ca64-4df7-9924-fbcd75b38d86-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-stmd9\" (UID: \"00da1369-ca64-4df7-9924-fbcd75b38d86\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-stmd9" Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.645269 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00da1369-ca64-4df7-9924-fbcd75b38d86-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-stmd9\" (UID: \"00da1369-ca64-4df7-9924-fbcd75b38d86\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-stmd9" Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.645349 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00da1369-ca64-4df7-9924-fbcd75b38d86-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-stmd9\" (UID: \"00da1369-ca64-4df7-9924-fbcd75b38d86\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-stmd9" Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.645456 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dqtq\" (UniqueName: \"kubernetes.io/projected/00da1369-ca64-4df7-9924-fbcd75b38d86-kube-api-access-6dqtq\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-stmd9\" (UID: \"00da1369-ca64-4df7-9924-fbcd75b38d86\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-stmd9" Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.650219 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00da1369-ca64-4df7-9924-fbcd75b38d86-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-stmd9\" (UID: \"00da1369-ca64-4df7-9924-fbcd75b38d86\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-stmd9" Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.661113 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00da1369-ca64-4df7-9924-fbcd75b38d86-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-stmd9\" (UID: \"00da1369-ca64-4df7-9924-fbcd75b38d86\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-stmd9" Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.665508 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dqtq\" (UniqueName: \"kubernetes.io/projected/00da1369-ca64-4df7-9924-fbcd75b38d86-kube-api-access-6dqtq\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-stmd9\" (UID: \"00da1369-ca64-4df7-9924-fbcd75b38d86\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-stmd9" Dec 09 10:25:42 crc kubenswrapper[4824]: I1209 10:25:42.812984 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-stmd9" Dec 09 10:25:43 crc kubenswrapper[4824]: I1209 10:25:43.374728 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-stmd9"] Dec 09 10:25:44 crc kubenswrapper[4824]: I1209 10:25:44.373160 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-stmd9" event={"ID":"00da1369-ca64-4df7-9924-fbcd75b38d86","Type":"ContainerStarted","Data":"0bb3c3ad404cf32456a9190fd07e97e1cb39a906aee70dd9e4ca4c4b9658232e"} Dec 09 10:25:44 crc kubenswrapper[4824]: I1209 10:25:44.373501 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-stmd9" event={"ID":"00da1369-ca64-4df7-9924-fbcd75b38d86","Type":"ContainerStarted","Data":"136549b69acd18474de9fc1f3e718976169f5f7f9d1b34676e28c58613461f55"} Dec 09 10:25:44 crc kubenswrapper[4824]: I1209 10:25:44.400121 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-stmd9" podStartSLOduration=1.985820943 podStartE2EDuration="2.400093137s" podCreationTimestamp="2025-12-09 10:25:42 +0000 UTC" firstStartedPulling="2025-12-09 10:25:43.381207392 +0000 UTC m=+2299.715712059" lastFinishedPulling="2025-12-09 10:25:43.795479586 +0000 UTC m=+2300.129984253" observedRunningTime="2025-12-09 10:25:44.390271659 +0000 UTC m=+2300.724776326" watchObservedRunningTime="2025-12-09 10:25:44.400093137 +0000 UTC m=+2300.734597804" Dec 09 10:25:49 crc kubenswrapper[4824]: I1209 10:25:49.435902 4824 generic.go:334] "Generic (PLEG): container finished" podID="00da1369-ca64-4df7-9924-fbcd75b38d86" containerID="0bb3c3ad404cf32456a9190fd07e97e1cb39a906aee70dd9e4ca4c4b9658232e" exitCode=0 Dec 09 10:25:49 crc kubenswrapper[4824]: I1209 10:25:49.435969 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-stmd9" event={"ID":"00da1369-ca64-4df7-9924-fbcd75b38d86","Type":"ContainerDied","Data":"0bb3c3ad404cf32456a9190fd07e97e1cb39a906aee70dd9e4ca4c4b9658232e"} Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.010925 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-stmd9" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.051812 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00da1369-ca64-4df7-9924-fbcd75b38d86-ssh-key\") pod \"00da1369-ca64-4df7-9924-fbcd75b38d86\" (UID: \"00da1369-ca64-4df7-9924-fbcd75b38d86\") " Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.052012 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00da1369-ca64-4df7-9924-fbcd75b38d86-inventory\") pod \"00da1369-ca64-4df7-9924-fbcd75b38d86\" (UID: \"00da1369-ca64-4df7-9924-fbcd75b38d86\") " Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.052066 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dqtq\" (UniqueName: \"kubernetes.io/projected/00da1369-ca64-4df7-9924-fbcd75b38d86-kube-api-access-6dqtq\") pod \"00da1369-ca64-4df7-9924-fbcd75b38d86\" (UID: \"00da1369-ca64-4df7-9924-fbcd75b38d86\") " Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.064092 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00da1369-ca64-4df7-9924-fbcd75b38d86-kube-api-access-6dqtq" (OuterVolumeSpecName: "kube-api-access-6dqtq") pod "00da1369-ca64-4df7-9924-fbcd75b38d86" (UID: "00da1369-ca64-4df7-9924-fbcd75b38d86"). InnerVolumeSpecName "kube-api-access-6dqtq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.098095 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00da1369-ca64-4df7-9924-fbcd75b38d86-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "00da1369-ca64-4df7-9924-fbcd75b38d86" (UID: "00da1369-ca64-4df7-9924-fbcd75b38d86"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.101214 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00da1369-ca64-4df7-9924-fbcd75b38d86-inventory" (OuterVolumeSpecName: "inventory") pod "00da1369-ca64-4df7-9924-fbcd75b38d86" (UID: "00da1369-ca64-4df7-9924-fbcd75b38d86"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.155855 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00da1369-ca64-4df7-9924-fbcd75b38d86-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.155897 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00da1369-ca64-4df7-9924-fbcd75b38d86-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.155910 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dqtq\" (UniqueName: \"kubernetes.io/projected/00da1369-ca64-4df7-9924-fbcd75b38d86-kube-api-access-6dqtq\") on node \"crc\" DevicePath \"\"" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.476054 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-stmd9" event={"ID":"00da1369-ca64-4df7-9924-fbcd75b38d86","Type":"ContainerDied","Data":"136549b69acd18474de9fc1f3e718976169f5f7f9d1b34676e28c58613461f55"} Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.476095 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="136549b69acd18474de9fc1f3e718976169f5f7f9d1b34676e28c58613461f55" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.476168 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-stmd9" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.606924 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-7m2lk"] Dec 09 10:25:51 crc kubenswrapper[4824]: E1209 10:25:51.607492 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00da1369-ca64-4df7-9924-fbcd75b38d86" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.607506 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="00da1369-ca64-4df7-9924-fbcd75b38d86" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.607797 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="00da1369-ca64-4df7-9924-fbcd75b38d86" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.608747 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7m2lk" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.612670 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.614212 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.614236 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.614678 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5nzld" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.669691 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a872b5e1-5510-4c90-bd4b-ebaf84dae414-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7m2lk\" (UID: \"a872b5e1-5510-4c90-bd4b-ebaf84dae414\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7m2lk" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.669762 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a872b5e1-5510-4c90-bd4b-ebaf84dae414-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7m2lk\" (UID: \"a872b5e1-5510-4c90-bd4b-ebaf84dae414\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7m2lk" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.669809 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q77mc\" (UniqueName: \"kubernetes.io/projected/a872b5e1-5510-4c90-bd4b-ebaf84dae414-kube-api-access-q77mc\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7m2lk\" (UID: \"a872b5e1-5510-4c90-bd4b-ebaf84dae414\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7m2lk" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.691234 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-7m2lk"] Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.771273 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q77mc\" (UniqueName: \"kubernetes.io/projected/a872b5e1-5510-4c90-bd4b-ebaf84dae414-kube-api-access-q77mc\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7m2lk\" (UID: \"a872b5e1-5510-4c90-bd4b-ebaf84dae414\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7m2lk" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.771495 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a872b5e1-5510-4c90-bd4b-ebaf84dae414-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7m2lk\" (UID: \"a872b5e1-5510-4c90-bd4b-ebaf84dae414\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7m2lk" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.771524 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a872b5e1-5510-4c90-bd4b-ebaf84dae414-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7m2lk\" (UID: \"a872b5e1-5510-4c90-bd4b-ebaf84dae414\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7m2lk" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.781593 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a872b5e1-5510-4c90-bd4b-ebaf84dae414-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7m2lk\" (UID: \"a872b5e1-5510-4c90-bd4b-ebaf84dae414\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7m2lk" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.783352 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a872b5e1-5510-4c90-bd4b-ebaf84dae414-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7m2lk\" (UID: \"a872b5e1-5510-4c90-bd4b-ebaf84dae414\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7m2lk" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.808452 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q77mc\" (UniqueName: \"kubernetes.io/projected/a872b5e1-5510-4c90-bd4b-ebaf84dae414-kube-api-access-q77mc\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7m2lk\" (UID: \"a872b5e1-5510-4c90-bd4b-ebaf84dae414\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7m2lk" Dec 09 10:25:51 crc kubenswrapper[4824]: I1209 10:25:51.992403 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7m2lk" Dec 09 10:25:52 crc kubenswrapper[4824]: I1209 10:25:52.578106 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-7m2lk"] Dec 09 10:25:53 crc kubenswrapper[4824]: I1209 10:25:53.500087 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7m2lk" event={"ID":"a872b5e1-5510-4c90-bd4b-ebaf84dae414","Type":"ContainerStarted","Data":"08047b4867790ab2cfb98c23e4f4491753703c505d5b56144f5ca5d154668a58"} Dec 09 10:25:54 crc kubenswrapper[4824]: I1209 10:25:54.522338 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7m2lk" event={"ID":"a872b5e1-5510-4c90-bd4b-ebaf84dae414","Type":"ContainerStarted","Data":"6793fcda8e5aff22874db9de88f26a94c042f906977109632e91d31239d8be9f"} Dec 09 10:25:54 crc kubenswrapper[4824]: I1209 10:25:54.550406 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7m2lk" podStartSLOduration=2.796780388 podStartE2EDuration="3.55038158s" podCreationTimestamp="2025-12-09 10:25:51 +0000 UTC" firstStartedPulling="2025-12-09 10:25:52.589018172 +0000 UTC m=+2308.923522839" lastFinishedPulling="2025-12-09 10:25:53.342619364 +0000 UTC m=+2309.677124031" observedRunningTime="2025-12-09 10:25:54.541048628 +0000 UTC m=+2310.875553285" watchObservedRunningTime="2025-12-09 10:25:54.55038158 +0000 UTC m=+2310.884886247" Dec 09 10:26:17 crc kubenswrapper[4824]: I1209 10:26:17.287549 4824 scope.go:117] "RemoveContainer" containerID="f5d20ae0706a64448caf4d1146e14931c247faafec1c1471ff9e5c97c20a91a4" Dec 09 10:26:32 crc kubenswrapper[4824]: I1209 10:26:32.860854 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:26:32 crc kubenswrapper[4824]: I1209 10:26:32.861372 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:26:33 crc kubenswrapper[4824]: I1209 10:26:33.982164 4824 generic.go:334] "Generic (PLEG): container finished" podID="a872b5e1-5510-4c90-bd4b-ebaf84dae414" containerID="6793fcda8e5aff22874db9de88f26a94c042f906977109632e91d31239d8be9f" exitCode=0 Dec 09 10:26:33 crc kubenswrapper[4824]: I1209 10:26:33.982261 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7m2lk" event={"ID":"a872b5e1-5510-4c90-bd4b-ebaf84dae414","Type":"ContainerDied","Data":"6793fcda8e5aff22874db9de88f26a94c042f906977109632e91d31239d8be9f"} Dec 09 10:26:35 crc kubenswrapper[4824]: I1209 10:26:35.484343 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7m2lk" Dec 09 10:26:35 crc kubenswrapper[4824]: I1209 10:26:35.605082 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a872b5e1-5510-4c90-bd4b-ebaf84dae414-inventory\") pod \"a872b5e1-5510-4c90-bd4b-ebaf84dae414\" (UID: \"a872b5e1-5510-4c90-bd4b-ebaf84dae414\") " Dec 09 10:26:35 crc kubenswrapper[4824]: I1209 10:26:35.605128 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a872b5e1-5510-4c90-bd4b-ebaf84dae414-ssh-key\") pod \"a872b5e1-5510-4c90-bd4b-ebaf84dae414\" (UID: \"a872b5e1-5510-4c90-bd4b-ebaf84dae414\") " Dec 09 10:26:35 crc kubenswrapper[4824]: I1209 10:26:35.605198 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q77mc\" (UniqueName: \"kubernetes.io/projected/a872b5e1-5510-4c90-bd4b-ebaf84dae414-kube-api-access-q77mc\") pod \"a872b5e1-5510-4c90-bd4b-ebaf84dae414\" (UID: \"a872b5e1-5510-4c90-bd4b-ebaf84dae414\") " Dec 09 10:26:35 crc kubenswrapper[4824]: I1209 10:26:35.611251 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a872b5e1-5510-4c90-bd4b-ebaf84dae414-kube-api-access-q77mc" (OuterVolumeSpecName: "kube-api-access-q77mc") pod "a872b5e1-5510-4c90-bd4b-ebaf84dae414" (UID: "a872b5e1-5510-4c90-bd4b-ebaf84dae414"). InnerVolumeSpecName "kube-api-access-q77mc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:26:35 crc kubenswrapper[4824]: I1209 10:26:35.638819 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a872b5e1-5510-4c90-bd4b-ebaf84dae414-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a872b5e1-5510-4c90-bd4b-ebaf84dae414" (UID: "a872b5e1-5510-4c90-bd4b-ebaf84dae414"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:26:35 crc kubenswrapper[4824]: I1209 10:26:35.642645 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a872b5e1-5510-4c90-bd4b-ebaf84dae414-inventory" (OuterVolumeSpecName: "inventory") pod "a872b5e1-5510-4c90-bd4b-ebaf84dae414" (UID: "a872b5e1-5510-4c90-bd4b-ebaf84dae414"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:26:35 crc kubenswrapper[4824]: I1209 10:26:35.708101 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a872b5e1-5510-4c90-bd4b-ebaf84dae414-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 10:26:35 crc kubenswrapper[4824]: I1209 10:26:35.708141 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a872b5e1-5510-4c90-bd4b-ebaf84dae414-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 10:26:35 crc kubenswrapper[4824]: I1209 10:26:35.708151 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q77mc\" (UniqueName: \"kubernetes.io/projected/a872b5e1-5510-4c90-bd4b-ebaf84dae414-kube-api-access-q77mc\") on node \"crc\" DevicePath \"\"" Dec 09 10:26:36 crc kubenswrapper[4824]: I1209 10:26:36.005519 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7m2lk" event={"ID":"a872b5e1-5510-4c90-bd4b-ebaf84dae414","Type":"ContainerDied","Data":"08047b4867790ab2cfb98c23e4f4491753703c505d5b56144f5ca5d154668a58"} Dec 09 10:26:36 crc kubenswrapper[4824]: I1209 10:26:36.005879 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08047b4867790ab2cfb98c23e4f4491753703c505d5b56144f5ca5d154668a58" Dec 09 10:26:36 crc kubenswrapper[4824]: I1209 10:26:36.005565 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7m2lk" Dec 09 10:26:36 crc kubenswrapper[4824]: I1209 10:26:36.100771 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs"] Dec 09 10:26:36 crc kubenswrapper[4824]: E1209 10:26:36.101386 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a872b5e1-5510-4c90-bd4b-ebaf84dae414" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 09 10:26:36 crc kubenswrapper[4824]: I1209 10:26:36.101408 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a872b5e1-5510-4c90-bd4b-ebaf84dae414" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 09 10:26:36 crc kubenswrapper[4824]: I1209 10:26:36.101621 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a872b5e1-5510-4c90-bd4b-ebaf84dae414" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 09 10:26:36 crc kubenswrapper[4824]: I1209 10:26:36.102581 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs" Dec 09 10:26:36 crc kubenswrapper[4824]: I1209 10:26:36.104415 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 10:26:36 crc kubenswrapper[4824]: I1209 10:26:36.105021 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 10:26:36 crc kubenswrapper[4824]: I1209 10:26:36.105335 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 10:26:36 crc kubenswrapper[4824]: I1209 10:26:36.105499 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5nzld" Dec 09 10:26:36 crc kubenswrapper[4824]: I1209 10:26:36.114748 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs"] Dec 09 10:26:36 crc kubenswrapper[4824]: I1209 10:26:36.220555 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/721d01dd-ea5d-41e9-a01c-c6fe4ba28086-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"721d01dd-ea5d-41e9-a01c-c6fe4ba28086\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs" Dec 09 10:26:36 crc kubenswrapper[4824]: I1209 10:26:36.220692 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/721d01dd-ea5d-41e9-a01c-c6fe4ba28086-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"721d01dd-ea5d-41e9-a01c-c6fe4ba28086\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs" Dec 09 10:26:36 crc kubenswrapper[4824]: I1209 10:26:36.220828 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drtq7\" (UniqueName: \"kubernetes.io/projected/721d01dd-ea5d-41e9-a01c-c6fe4ba28086-kube-api-access-drtq7\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"721d01dd-ea5d-41e9-a01c-c6fe4ba28086\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs" Dec 09 10:26:36 crc kubenswrapper[4824]: I1209 10:26:36.322830 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/721d01dd-ea5d-41e9-a01c-c6fe4ba28086-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"721d01dd-ea5d-41e9-a01c-c6fe4ba28086\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs" Dec 09 10:26:36 crc kubenswrapper[4824]: I1209 10:26:36.322950 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/721d01dd-ea5d-41e9-a01c-c6fe4ba28086-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"721d01dd-ea5d-41e9-a01c-c6fe4ba28086\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs" Dec 09 10:26:36 crc kubenswrapper[4824]: I1209 10:26:36.323041 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drtq7\" (UniqueName: \"kubernetes.io/projected/721d01dd-ea5d-41e9-a01c-c6fe4ba28086-kube-api-access-drtq7\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"721d01dd-ea5d-41e9-a01c-c6fe4ba28086\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs" Dec 09 10:26:36 crc kubenswrapper[4824]: I1209 10:26:36.327515 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/721d01dd-ea5d-41e9-a01c-c6fe4ba28086-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"721d01dd-ea5d-41e9-a01c-c6fe4ba28086\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs" Dec 09 10:26:36 crc kubenswrapper[4824]: I1209 10:26:36.331592 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/721d01dd-ea5d-41e9-a01c-c6fe4ba28086-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"721d01dd-ea5d-41e9-a01c-c6fe4ba28086\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs" Dec 09 10:26:36 crc kubenswrapper[4824]: I1209 10:26:36.338564 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drtq7\" (UniqueName: \"kubernetes.io/projected/721d01dd-ea5d-41e9-a01c-c6fe4ba28086-kube-api-access-drtq7\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"721d01dd-ea5d-41e9-a01c-c6fe4ba28086\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs" Dec 09 10:26:36 crc kubenswrapper[4824]: I1209 10:26:36.424950 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs" Dec 09 10:26:36 crc kubenswrapper[4824]: I1209 10:26:36.992658 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs"] Dec 09 10:26:37 crc kubenswrapper[4824]: I1209 10:26:37.016926 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs" event={"ID":"721d01dd-ea5d-41e9-a01c-c6fe4ba28086","Type":"ContainerStarted","Data":"0edd8e6dd98b8bf458e609573735dacadef0fd248b1a738df1b8cb9cc9398049"} Dec 09 10:26:38 crc kubenswrapper[4824]: I1209 10:26:38.027680 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs" event={"ID":"721d01dd-ea5d-41e9-a01c-c6fe4ba28086","Type":"ContainerStarted","Data":"50ca3b7ae002dd9b2d56f12aedd21bd15ee24fbf14ef74039357b21d9c8d4089"} Dec 09 10:26:38 crc kubenswrapper[4824]: I1209 10:26:38.051087 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs" podStartSLOduration=1.6352955439999999 podStartE2EDuration="2.051053584s" podCreationTimestamp="2025-12-09 10:26:36 +0000 UTC" firstStartedPulling="2025-12-09 10:26:36.995733089 +0000 UTC m=+2353.330237756" lastFinishedPulling="2025-12-09 10:26:37.411491129 +0000 UTC m=+2353.745995796" observedRunningTime="2025-12-09 10:26:38.048224405 +0000 UTC m=+2354.382729092" watchObservedRunningTime="2025-12-09 10:26:38.051053584 +0000 UTC m=+2354.385558251" Dec 09 10:27:02 crc kubenswrapper[4824]: I1209 10:27:02.861001 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:27:02 crc kubenswrapper[4824]: I1209 10:27:02.861575 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:27:25 crc kubenswrapper[4824]: I1209 10:27:25.061125 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-m6t59"] Dec 09 10:27:25 crc kubenswrapper[4824]: I1209 10:27:25.074913 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-m6t59"] Dec 09 10:27:25 crc kubenswrapper[4824]: I1209 10:27:25.941483 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8051e79f-a20f-4a10-b2f8-d8c770328797" path="/var/lib/kubelet/pods/8051e79f-a20f-4a10-b2f8-d8c770328797/volumes" Dec 09 10:27:30 crc kubenswrapper[4824]: I1209 10:27:30.681949 4824 generic.go:334] "Generic (PLEG): container finished" podID="721d01dd-ea5d-41e9-a01c-c6fe4ba28086" containerID="50ca3b7ae002dd9b2d56f12aedd21bd15ee24fbf14ef74039357b21d9c8d4089" exitCode=0 Dec 09 10:27:30 crc kubenswrapper[4824]: I1209 10:27:30.682082 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs" event={"ID":"721d01dd-ea5d-41e9-a01c-c6fe4ba28086","Type":"ContainerDied","Data":"50ca3b7ae002dd9b2d56f12aedd21bd15ee24fbf14ef74039357b21d9c8d4089"} Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.223391 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.403810 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drtq7\" (UniqueName: \"kubernetes.io/projected/721d01dd-ea5d-41e9-a01c-c6fe4ba28086-kube-api-access-drtq7\") pod \"721d01dd-ea5d-41e9-a01c-c6fe4ba28086\" (UID: \"721d01dd-ea5d-41e9-a01c-c6fe4ba28086\") " Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.404709 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/721d01dd-ea5d-41e9-a01c-c6fe4ba28086-inventory\") pod \"721d01dd-ea5d-41e9-a01c-c6fe4ba28086\" (UID: \"721d01dd-ea5d-41e9-a01c-c6fe4ba28086\") " Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.404973 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/721d01dd-ea5d-41e9-a01c-c6fe4ba28086-ssh-key\") pod \"721d01dd-ea5d-41e9-a01c-c6fe4ba28086\" (UID: \"721d01dd-ea5d-41e9-a01c-c6fe4ba28086\") " Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.412121 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/721d01dd-ea5d-41e9-a01c-c6fe4ba28086-kube-api-access-drtq7" (OuterVolumeSpecName: "kube-api-access-drtq7") pod "721d01dd-ea5d-41e9-a01c-c6fe4ba28086" (UID: "721d01dd-ea5d-41e9-a01c-c6fe4ba28086"). InnerVolumeSpecName "kube-api-access-drtq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.439983 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/721d01dd-ea5d-41e9-a01c-c6fe4ba28086-inventory" (OuterVolumeSpecName: "inventory") pod "721d01dd-ea5d-41e9-a01c-c6fe4ba28086" (UID: "721d01dd-ea5d-41e9-a01c-c6fe4ba28086"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.458395 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/721d01dd-ea5d-41e9-a01c-c6fe4ba28086-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "721d01dd-ea5d-41e9-a01c-c6fe4ba28086" (UID: "721d01dd-ea5d-41e9-a01c-c6fe4ba28086"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.510592 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/721d01dd-ea5d-41e9-a01c-c6fe4ba28086-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.510635 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/721d01dd-ea5d-41e9-a01c-c6fe4ba28086-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.510644 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drtq7\" (UniqueName: \"kubernetes.io/projected/721d01dd-ea5d-41e9-a01c-c6fe4ba28086-kube-api-access-drtq7\") on node \"crc\" DevicePath \"\"" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.705450 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs" event={"ID":"721d01dd-ea5d-41e9-a01c-c6fe4ba28086","Type":"ContainerDied","Data":"0edd8e6dd98b8bf458e609573735dacadef0fd248b1a738df1b8cb9cc9398049"} Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.705517 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0edd8e6dd98b8bf458e609573735dacadef0fd248b1a738df1b8cb9cc9398049" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.705521 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.804257 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-6tlwx"] Dec 09 10:27:32 crc kubenswrapper[4824]: E1209 10:27:32.804985 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="721d01dd-ea5d-41e9-a01c-c6fe4ba28086" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.805008 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="721d01dd-ea5d-41e9-a01c-c6fe4ba28086" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.805357 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="721d01dd-ea5d-41e9-a01c-c6fe4ba28086" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.806343 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-6tlwx" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.809022 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5nzld" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.809189 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.809263 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.809325 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.829902 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/eae91833-e890-428b-beee-3420b3718f45-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-6tlwx\" (UID: \"eae91833-e890-428b-beee-3420b3718f45\") " pod="openstack/ssh-known-hosts-edpm-deployment-6tlwx" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.830068 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-6tlwx"] Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.830953 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/eae91833-e890-428b-beee-3420b3718f45-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-6tlwx\" (UID: \"eae91833-e890-428b-beee-3420b3718f45\") " pod="openstack/ssh-known-hosts-edpm-deployment-6tlwx" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.861153 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.861238 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.861290 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.862378 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008"} pod="openshift-machine-config-operator/machine-config-daemon-dth8x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.862452 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" containerID="cri-o://2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" gracePeriod=600 Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.933792 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4nhp\" (UniqueName: \"kubernetes.io/projected/eae91833-e890-428b-beee-3420b3718f45-kube-api-access-t4nhp\") pod \"ssh-known-hosts-edpm-deployment-6tlwx\" (UID: \"eae91833-e890-428b-beee-3420b3718f45\") " pod="openstack/ssh-known-hosts-edpm-deployment-6tlwx" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.933972 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/eae91833-e890-428b-beee-3420b3718f45-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-6tlwx\" (UID: \"eae91833-e890-428b-beee-3420b3718f45\") " pod="openstack/ssh-known-hosts-edpm-deployment-6tlwx" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.934931 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/eae91833-e890-428b-beee-3420b3718f45-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-6tlwx\" (UID: \"eae91833-e890-428b-beee-3420b3718f45\") " pod="openstack/ssh-known-hosts-edpm-deployment-6tlwx" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.937916 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/eae91833-e890-428b-beee-3420b3718f45-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-6tlwx\" (UID: \"eae91833-e890-428b-beee-3420b3718f45\") " pod="openstack/ssh-known-hosts-edpm-deployment-6tlwx" Dec 09 10:27:32 crc kubenswrapper[4824]: I1209 10:27:32.949584 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/eae91833-e890-428b-beee-3420b3718f45-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-6tlwx\" (UID: \"eae91833-e890-428b-beee-3420b3718f45\") " pod="openstack/ssh-known-hosts-edpm-deployment-6tlwx" Dec 09 10:27:32 crc kubenswrapper[4824]: E1209 10:27:32.991094 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:27:33 crc kubenswrapper[4824]: I1209 10:27:33.039273 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4nhp\" (UniqueName: \"kubernetes.io/projected/eae91833-e890-428b-beee-3420b3718f45-kube-api-access-t4nhp\") pod \"ssh-known-hosts-edpm-deployment-6tlwx\" (UID: \"eae91833-e890-428b-beee-3420b3718f45\") " pod="openstack/ssh-known-hosts-edpm-deployment-6tlwx" Dec 09 10:27:33 crc kubenswrapper[4824]: I1209 10:27:33.062413 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4nhp\" (UniqueName: \"kubernetes.io/projected/eae91833-e890-428b-beee-3420b3718f45-kube-api-access-t4nhp\") pod \"ssh-known-hosts-edpm-deployment-6tlwx\" (UID: \"eae91833-e890-428b-beee-3420b3718f45\") " pod="openstack/ssh-known-hosts-edpm-deployment-6tlwx" Dec 09 10:27:33 crc kubenswrapper[4824]: I1209 10:27:33.127217 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-6tlwx" Dec 09 10:27:33 crc kubenswrapper[4824]: I1209 10:27:33.658891 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-6tlwx"] Dec 09 10:27:33 crc kubenswrapper[4824]: I1209 10:27:33.731614 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-6tlwx" event={"ID":"eae91833-e890-428b-beee-3420b3718f45","Type":"ContainerStarted","Data":"65f3f4fb907a03d88a93de4792cd4f1638545c508dfcdfd31451871533b83c5a"} Dec 09 10:27:33 crc kubenswrapper[4824]: I1209 10:27:33.734844 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" exitCode=0 Dec 09 10:27:33 crc kubenswrapper[4824]: I1209 10:27:33.734889 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerDied","Data":"2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008"} Dec 09 10:27:33 crc kubenswrapper[4824]: I1209 10:27:33.734930 4824 scope.go:117] "RemoveContainer" containerID="d4d2394221cb52bd0aff000b3ed8a1bb39ce26a5dee3baa44cc5e9dfefc07899" Dec 09 10:27:33 crc kubenswrapper[4824]: I1209 10:27:33.736895 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:27:33 crc kubenswrapper[4824]: E1209 10:27:33.737417 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:27:34 crc kubenswrapper[4824]: I1209 10:27:34.751586 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-6tlwx" event={"ID":"eae91833-e890-428b-beee-3420b3718f45","Type":"ContainerStarted","Data":"f1b8ce345fdab3d00191c7fac44ac39d767aab216693d089bc30e33853eea384"} Dec 09 10:27:34 crc kubenswrapper[4824]: I1209 10:27:34.787311 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-6tlwx" podStartSLOduration=2.337529688 podStartE2EDuration="2.787240251s" podCreationTimestamp="2025-12-09 10:27:32 +0000 UTC" firstStartedPulling="2025-12-09 10:27:33.664410463 +0000 UTC m=+2409.998915120" lastFinishedPulling="2025-12-09 10:27:34.114121016 +0000 UTC m=+2410.448625683" observedRunningTime="2025-12-09 10:27:34.776290379 +0000 UTC m=+2411.110795096" watchObservedRunningTime="2025-12-09 10:27:34.787240251 +0000 UTC m=+2411.121744928" Dec 09 10:27:41 crc kubenswrapper[4824]: I1209 10:27:41.880087 4824 generic.go:334] "Generic (PLEG): container finished" podID="eae91833-e890-428b-beee-3420b3718f45" containerID="f1b8ce345fdab3d00191c7fac44ac39d767aab216693d089bc30e33853eea384" exitCode=0 Dec 09 10:27:41 crc kubenswrapper[4824]: I1209 10:27:41.880208 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-6tlwx" event={"ID":"eae91833-e890-428b-beee-3420b3718f45","Type":"ContainerDied","Data":"f1b8ce345fdab3d00191c7fac44ac39d767aab216693d089bc30e33853eea384"} Dec 09 10:27:43 crc kubenswrapper[4824]: I1209 10:27:43.461671 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-6tlwx" Dec 09 10:27:43 crc kubenswrapper[4824]: I1209 10:27:43.654745 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/eae91833-e890-428b-beee-3420b3718f45-ssh-key-openstack-edpm-ipam\") pod \"eae91833-e890-428b-beee-3420b3718f45\" (UID: \"eae91833-e890-428b-beee-3420b3718f45\") " Dec 09 10:27:43 crc kubenswrapper[4824]: I1209 10:27:43.654959 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/eae91833-e890-428b-beee-3420b3718f45-inventory-0\") pod \"eae91833-e890-428b-beee-3420b3718f45\" (UID: \"eae91833-e890-428b-beee-3420b3718f45\") " Dec 09 10:27:43 crc kubenswrapper[4824]: I1209 10:27:43.655068 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4nhp\" (UniqueName: \"kubernetes.io/projected/eae91833-e890-428b-beee-3420b3718f45-kube-api-access-t4nhp\") pod \"eae91833-e890-428b-beee-3420b3718f45\" (UID: \"eae91833-e890-428b-beee-3420b3718f45\") " Dec 09 10:27:43 crc kubenswrapper[4824]: I1209 10:27:43.660677 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eae91833-e890-428b-beee-3420b3718f45-kube-api-access-t4nhp" (OuterVolumeSpecName: "kube-api-access-t4nhp") pod "eae91833-e890-428b-beee-3420b3718f45" (UID: "eae91833-e890-428b-beee-3420b3718f45"). InnerVolumeSpecName "kube-api-access-t4nhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:27:43 crc kubenswrapper[4824]: I1209 10:27:43.686158 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eae91833-e890-428b-beee-3420b3718f45-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "eae91833-e890-428b-beee-3420b3718f45" (UID: "eae91833-e890-428b-beee-3420b3718f45"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:27:43 crc kubenswrapper[4824]: I1209 10:27:43.694196 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eae91833-e890-428b-beee-3420b3718f45-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "eae91833-e890-428b-beee-3420b3718f45" (UID: "eae91833-e890-428b-beee-3420b3718f45"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:27:43 crc kubenswrapper[4824]: I1209 10:27:43.758597 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/eae91833-e890-428b-beee-3420b3718f45-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 09 10:27:43 crc kubenswrapper[4824]: I1209 10:27:43.758637 4824 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/eae91833-e890-428b-beee-3420b3718f45-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:27:43 crc kubenswrapper[4824]: I1209 10:27:43.758648 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4nhp\" (UniqueName: \"kubernetes.io/projected/eae91833-e890-428b-beee-3420b3718f45-kube-api-access-t4nhp\") on node \"crc\" DevicePath \"\"" Dec 09 10:27:43 crc kubenswrapper[4824]: I1209 10:27:43.923113 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-6tlwx" Dec 09 10:27:43 crc kubenswrapper[4824]: I1209 10:27:43.942603 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-6tlwx" event={"ID":"eae91833-e890-428b-beee-3420b3718f45","Type":"ContainerDied","Data":"65f3f4fb907a03d88a93de4792cd4f1638545c508dfcdfd31451871533b83c5a"} Dec 09 10:27:43 crc kubenswrapper[4824]: I1209 10:27:43.942642 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65f3f4fb907a03d88a93de4792cd4f1638545c508dfcdfd31451871533b83c5a" Dec 09 10:27:44 crc kubenswrapper[4824]: I1209 10:27:44.013761 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-d5rkl"] Dec 09 10:27:44 crc kubenswrapper[4824]: E1209 10:27:44.014528 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae91833-e890-428b-beee-3420b3718f45" containerName="ssh-known-hosts-edpm-deployment" Dec 09 10:27:44 crc kubenswrapper[4824]: I1209 10:27:44.014551 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae91833-e890-428b-beee-3420b3718f45" containerName="ssh-known-hosts-edpm-deployment" Dec 09 10:27:44 crc kubenswrapper[4824]: I1209 10:27:44.014881 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="eae91833-e890-428b-beee-3420b3718f45" containerName="ssh-known-hosts-edpm-deployment" Dec 09 10:27:44 crc kubenswrapper[4824]: I1209 10:27:44.016051 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d5rkl" Dec 09 10:27:44 crc kubenswrapper[4824]: I1209 10:27:44.019058 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 10:27:44 crc kubenswrapper[4824]: I1209 10:27:44.019314 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 10:27:44 crc kubenswrapper[4824]: I1209 10:27:44.019426 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 10:27:44 crc kubenswrapper[4824]: I1209 10:27:44.020203 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5nzld" Dec 09 10:27:44 crc kubenswrapper[4824]: I1209 10:27:44.039299 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-d5rkl"] Dec 09 10:27:44 crc kubenswrapper[4824]: I1209 10:27:44.067005 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3329105-6b83-49ff-8dd7-5343a656f7af-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d5rkl\" (UID: \"b3329105-6b83-49ff-8dd7-5343a656f7af\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d5rkl" Dec 09 10:27:44 crc kubenswrapper[4824]: I1209 10:27:44.067073 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3329105-6b83-49ff-8dd7-5343a656f7af-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d5rkl\" (UID: \"b3329105-6b83-49ff-8dd7-5343a656f7af\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d5rkl" Dec 09 10:27:44 crc kubenswrapper[4824]: I1209 10:27:44.067166 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tfp2\" (UniqueName: \"kubernetes.io/projected/b3329105-6b83-49ff-8dd7-5343a656f7af-kube-api-access-9tfp2\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d5rkl\" (UID: \"b3329105-6b83-49ff-8dd7-5343a656f7af\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d5rkl" Dec 09 10:27:44 crc kubenswrapper[4824]: I1209 10:27:44.170177 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3329105-6b83-49ff-8dd7-5343a656f7af-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d5rkl\" (UID: \"b3329105-6b83-49ff-8dd7-5343a656f7af\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d5rkl" Dec 09 10:27:44 crc kubenswrapper[4824]: I1209 10:27:44.170242 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3329105-6b83-49ff-8dd7-5343a656f7af-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d5rkl\" (UID: \"b3329105-6b83-49ff-8dd7-5343a656f7af\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d5rkl" Dec 09 10:27:44 crc kubenswrapper[4824]: I1209 10:27:44.170336 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tfp2\" (UniqueName: \"kubernetes.io/projected/b3329105-6b83-49ff-8dd7-5343a656f7af-kube-api-access-9tfp2\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d5rkl\" (UID: \"b3329105-6b83-49ff-8dd7-5343a656f7af\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d5rkl" Dec 09 10:27:44 crc kubenswrapper[4824]: I1209 10:27:44.175265 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3329105-6b83-49ff-8dd7-5343a656f7af-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d5rkl\" (UID: \"b3329105-6b83-49ff-8dd7-5343a656f7af\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d5rkl" Dec 09 10:27:44 crc kubenswrapper[4824]: I1209 10:27:44.180339 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3329105-6b83-49ff-8dd7-5343a656f7af-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d5rkl\" (UID: \"b3329105-6b83-49ff-8dd7-5343a656f7af\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d5rkl" Dec 09 10:27:44 crc kubenswrapper[4824]: I1209 10:27:44.191032 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tfp2\" (UniqueName: \"kubernetes.io/projected/b3329105-6b83-49ff-8dd7-5343a656f7af-kube-api-access-9tfp2\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d5rkl\" (UID: \"b3329105-6b83-49ff-8dd7-5343a656f7af\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d5rkl" Dec 09 10:27:44 crc kubenswrapper[4824]: I1209 10:27:44.333963 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d5rkl" Dec 09 10:27:44 crc kubenswrapper[4824]: I1209 10:27:44.951410 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-d5rkl"] Dec 09 10:27:45 crc kubenswrapper[4824]: I1209 10:27:45.952506 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d5rkl" event={"ID":"b3329105-6b83-49ff-8dd7-5343a656f7af","Type":"ContainerStarted","Data":"c9aca7c130f1dccde7a75eca047b362718d1813821f23b706678ed4cc55a885f"} Dec 09 10:27:45 crc kubenswrapper[4824]: I1209 10:27:45.953156 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d5rkl" event={"ID":"b3329105-6b83-49ff-8dd7-5343a656f7af","Type":"ContainerStarted","Data":"6afc5f7e3dc6ca26ba854a2d03f84f95d095a55deab9ed4ad1a5217505397488"} Dec 09 10:27:45 crc kubenswrapper[4824]: I1209 10:27:45.984559 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d5rkl" podStartSLOduration=2.464005291 podStartE2EDuration="2.98453265s" podCreationTimestamp="2025-12-09 10:27:43 +0000 UTC" firstStartedPulling="2025-12-09 10:27:44.947672572 +0000 UTC m=+2421.282177239" lastFinishedPulling="2025-12-09 10:27:45.468199931 +0000 UTC m=+2421.802704598" observedRunningTime="2025-12-09 10:27:45.970805071 +0000 UTC m=+2422.305309728" watchObservedRunningTime="2025-12-09 10:27:45.98453265 +0000 UTC m=+2422.319037317" Dec 09 10:27:47 crc kubenswrapper[4824]: I1209 10:27:47.912714 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:27:47 crc kubenswrapper[4824]: E1209 10:27:47.913635 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:27:54 crc kubenswrapper[4824]: I1209 10:27:54.042469 4824 generic.go:334] "Generic (PLEG): container finished" podID="b3329105-6b83-49ff-8dd7-5343a656f7af" containerID="c9aca7c130f1dccde7a75eca047b362718d1813821f23b706678ed4cc55a885f" exitCode=0 Dec 09 10:27:54 crc kubenswrapper[4824]: I1209 10:27:54.042582 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d5rkl" event={"ID":"b3329105-6b83-49ff-8dd7-5343a656f7af","Type":"ContainerDied","Data":"c9aca7c130f1dccde7a75eca047b362718d1813821f23b706678ed4cc55a885f"} Dec 09 10:27:55 crc kubenswrapper[4824]: I1209 10:27:55.703918 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d5rkl" Dec 09 10:27:55 crc kubenswrapper[4824]: I1209 10:27:55.896351 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3329105-6b83-49ff-8dd7-5343a656f7af-inventory\") pod \"b3329105-6b83-49ff-8dd7-5343a656f7af\" (UID: \"b3329105-6b83-49ff-8dd7-5343a656f7af\") " Dec 09 10:27:55 crc kubenswrapper[4824]: I1209 10:27:55.896933 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3329105-6b83-49ff-8dd7-5343a656f7af-ssh-key\") pod \"b3329105-6b83-49ff-8dd7-5343a656f7af\" (UID: \"b3329105-6b83-49ff-8dd7-5343a656f7af\") " Dec 09 10:27:55 crc kubenswrapper[4824]: I1209 10:27:55.897135 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9tfp2\" (UniqueName: \"kubernetes.io/projected/b3329105-6b83-49ff-8dd7-5343a656f7af-kube-api-access-9tfp2\") pod \"b3329105-6b83-49ff-8dd7-5343a656f7af\" (UID: \"b3329105-6b83-49ff-8dd7-5343a656f7af\") " Dec 09 10:27:55 crc kubenswrapper[4824]: I1209 10:27:55.904694 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3329105-6b83-49ff-8dd7-5343a656f7af-kube-api-access-9tfp2" (OuterVolumeSpecName: "kube-api-access-9tfp2") pod "b3329105-6b83-49ff-8dd7-5343a656f7af" (UID: "b3329105-6b83-49ff-8dd7-5343a656f7af"). InnerVolumeSpecName "kube-api-access-9tfp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:27:55 crc kubenswrapper[4824]: I1209 10:27:55.944290 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3329105-6b83-49ff-8dd7-5343a656f7af-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b3329105-6b83-49ff-8dd7-5343a656f7af" (UID: "b3329105-6b83-49ff-8dd7-5343a656f7af"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:27:55 crc kubenswrapper[4824]: I1209 10:27:55.944693 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3329105-6b83-49ff-8dd7-5343a656f7af-inventory" (OuterVolumeSpecName: "inventory") pod "b3329105-6b83-49ff-8dd7-5343a656f7af" (UID: "b3329105-6b83-49ff-8dd7-5343a656f7af"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.001729 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3329105-6b83-49ff-8dd7-5343a656f7af-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.001762 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9tfp2\" (UniqueName: \"kubernetes.io/projected/b3329105-6b83-49ff-8dd7-5343a656f7af-kube-api-access-9tfp2\") on node \"crc\" DevicePath \"\"" Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.001773 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3329105-6b83-49ff-8dd7-5343a656f7af-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.065396 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d5rkl" event={"ID":"b3329105-6b83-49ff-8dd7-5343a656f7af","Type":"ContainerDied","Data":"6afc5f7e3dc6ca26ba854a2d03f84f95d095a55deab9ed4ad1a5217505397488"} Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.065444 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6afc5f7e3dc6ca26ba854a2d03f84f95d095a55deab9ed4ad1a5217505397488" Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.065744 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d5rkl" Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.139234 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6"] Dec 09 10:27:56 crc kubenswrapper[4824]: E1209 10:27:56.140025 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3329105-6b83-49ff-8dd7-5343a656f7af" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.140044 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3329105-6b83-49ff-8dd7-5343a656f7af" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.140317 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3329105-6b83-49ff-8dd7-5343a656f7af" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.141236 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6" Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.144077 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.144332 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.144524 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5nzld" Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.145088 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.152154 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6"] Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.307752 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45zdx\" (UniqueName: \"kubernetes.io/projected/21d8d8f6-2356-406f-a811-e867f326d758-kube-api-access-45zdx\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6\" (UID: \"21d8d8f6-2356-406f-a811-e867f326d758\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6" Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.307936 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21d8d8f6-2356-406f-a811-e867f326d758-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6\" (UID: \"21d8d8f6-2356-406f-a811-e867f326d758\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6" Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.308658 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21d8d8f6-2356-406f-a811-e867f326d758-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6\" (UID: \"21d8d8f6-2356-406f-a811-e867f326d758\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6" Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.410670 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21d8d8f6-2356-406f-a811-e867f326d758-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6\" (UID: \"21d8d8f6-2356-406f-a811-e867f326d758\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6" Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.410761 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21d8d8f6-2356-406f-a811-e867f326d758-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6\" (UID: \"21d8d8f6-2356-406f-a811-e867f326d758\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6" Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.410915 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45zdx\" (UniqueName: \"kubernetes.io/projected/21d8d8f6-2356-406f-a811-e867f326d758-kube-api-access-45zdx\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6\" (UID: \"21d8d8f6-2356-406f-a811-e867f326d758\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6" Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.423607 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21d8d8f6-2356-406f-a811-e867f326d758-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6\" (UID: \"21d8d8f6-2356-406f-a811-e867f326d758\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6" Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.423710 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21d8d8f6-2356-406f-a811-e867f326d758-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6\" (UID: \"21d8d8f6-2356-406f-a811-e867f326d758\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6" Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.434003 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45zdx\" (UniqueName: \"kubernetes.io/projected/21d8d8f6-2356-406f-a811-e867f326d758-kube-api-access-45zdx\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6\" (UID: \"21d8d8f6-2356-406f-a811-e867f326d758\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6" Dec 09 10:27:56 crc kubenswrapper[4824]: I1209 10:27:56.458201 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6" Dec 09 10:27:57 crc kubenswrapper[4824]: I1209 10:27:57.026051 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6"] Dec 09 10:27:57 crc kubenswrapper[4824]: I1209 10:27:57.077465 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6" event={"ID":"21d8d8f6-2356-406f-a811-e867f326d758","Type":"ContainerStarted","Data":"d0cd6db956bce5e37007d3139b086b6db89e28416d2b6a4a49bc35ba7e056816"} Dec 09 10:27:58 crc kubenswrapper[4824]: I1209 10:27:58.089567 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6" event={"ID":"21d8d8f6-2356-406f-a811-e867f326d758","Type":"ContainerStarted","Data":"7e6186c52c6204c9343949d22f1146fece570f17920f956f887a093dbcaa6c45"} Dec 09 10:27:58 crc kubenswrapper[4824]: I1209 10:27:58.112714 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6" podStartSLOduration=1.600015535 podStartE2EDuration="2.112686839s" podCreationTimestamp="2025-12-09 10:27:56 +0000 UTC" firstStartedPulling="2025-12-09 10:27:57.026360433 +0000 UTC m=+2433.360865100" lastFinishedPulling="2025-12-09 10:27:57.539031727 +0000 UTC m=+2433.873536404" observedRunningTime="2025-12-09 10:27:58.103345036 +0000 UTC m=+2434.437849703" watchObservedRunningTime="2025-12-09 10:27:58.112686839 +0000 UTC m=+2434.447191516" Dec 09 10:27:59 crc kubenswrapper[4824]: I1209 10:27:59.912285 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:27:59 crc kubenswrapper[4824]: E1209 10:27:59.912858 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:28:05 crc kubenswrapper[4824]: I1209 10:28:05.046361 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-bpqmt"] Dec 09 10:28:05 crc kubenswrapper[4824]: I1209 10:28:05.058061 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-bpqmt"] Dec 09 10:28:05 crc kubenswrapper[4824]: I1209 10:28:05.923572 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e293dba-63b1-4841-b409-0271f0329251" path="/var/lib/kubelet/pods/3e293dba-63b1-4841-b409-0271f0329251/volumes" Dec 09 10:28:13 crc kubenswrapper[4824]: I1209 10:28:13.919341 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:28:13 crc kubenswrapper[4824]: E1209 10:28:13.920204 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:28:14 crc kubenswrapper[4824]: I1209 10:28:14.378815 4824 generic.go:334] "Generic (PLEG): container finished" podID="21d8d8f6-2356-406f-a811-e867f326d758" containerID="7e6186c52c6204c9343949d22f1146fece570f17920f956f887a093dbcaa6c45" exitCode=0 Dec 09 10:28:14 crc kubenswrapper[4824]: I1209 10:28:14.378858 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6" event={"ID":"21d8d8f6-2356-406f-a811-e867f326d758","Type":"ContainerDied","Data":"7e6186c52c6204c9343949d22f1146fece570f17920f956f887a093dbcaa6c45"} Dec 09 10:28:15 crc kubenswrapper[4824]: I1209 10:28:15.886964 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6" Dec 09 10:28:15 crc kubenswrapper[4824]: I1209 10:28:15.927403 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21d8d8f6-2356-406f-a811-e867f326d758-inventory\") pod \"21d8d8f6-2356-406f-a811-e867f326d758\" (UID: \"21d8d8f6-2356-406f-a811-e867f326d758\") " Dec 09 10:28:15 crc kubenswrapper[4824]: I1209 10:28:15.927557 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45zdx\" (UniqueName: \"kubernetes.io/projected/21d8d8f6-2356-406f-a811-e867f326d758-kube-api-access-45zdx\") pod \"21d8d8f6-2356-406f-a811-e867f326d758\" (UID: \"21d8d8f6-2356-406f-a811-e867f326d758\") " Dec 09 10:28:15 crc kubenswrapper[4824]: I1209 10:28:15.927646 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21d8d8f6-2356-406f-a811-e867f326d758-ssh-key\") pod \"21d8d8f6-2356-406f-a811-e867f326d758\" (UID: \"21d8d8f6-2356-406f-a811-e867f326d758\") " Dec 09 10:28:15 crc kubenswrapper[4824]: I1209 10:28:15.956556 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21d8d8f6-2356-406f-a811-e867f326d758-kube-api-access-45zdx" (OuterVolumeSpecName: "kube-api-access-45zdx") pod "21d8d8f6-2356-406f-a811-e867f326d758" (UID: "21d8d8f6-2356-406f-a811-e867f326d758"). InnerVolumeSpecName "kube-api-access-45zdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:28:15 crc kubenswrapper[4824]: I1209 10:28:15.962929 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21d8d8f6-2356-406f-a811-e867f326d758-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "21d8d8f6-2356-406f-a811-e867f326d758" (UID: "21d8d8f6-2356-406f-a811-e867f326d758"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:28:15 crc kubenswrapper[4824]: I1209 10:28:15.963545 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21d8d8f6-2356-406f-a811-e867f326d758-inventory" (OuterVolumeSpecName: "inventory") pod "21d8d8f6-2356-406f-a811-e867f326d758" (UID: "21d8d8f6-2356-406f-a811-e867f326d758"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.030540 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21d8d8f6-2356-406f-a811-e867f326d758-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.032244 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21d8d8f6-2356-406f-a811-e867f326d758-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.032382 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45zdx\" (UniqueName: \"kubernetes.io/projected/21d8d8f6-2356-406f-a811-e867f326d758-kube-api-access-45zdx\") on node \"crc\" DevicePath \"\"" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.400957 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6" event={"ID":"21d8d8f6-2356-406f-a811-e867f326d758","Type":"ContainerDied","Data":"d0cd6db956bce5e37007d3139b086b6db89e28416d2b6a4a49bc35ba7e056816"} Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.401270 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0cd6db956bce5e37007d3139b086b6db89e28416d2b6a4a49bc35ba7e056816" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.401408 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.504062 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts"] Dec 09 10:28:16 crc kubenswrapper[4824]: E1209 10:28:16.504661 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21d8d8f6-2356-406f-a811-e867f326d758" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.504678 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="21d8d8f6-2356-406f-a811-e867f326d758" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.504918 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="21d8d8f6-2356-406f-a811-e867f326d758" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.505910 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.508501 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.509012 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.509108 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.509530 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5nzld" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.509681 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.509694 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.509749 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.510074 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.510401 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.530668 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts"] Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.550572 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.550871 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.550977 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.551092 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.551140 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.551195 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qt76\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-kube-api-access-8qt76\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.551246 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.551285 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.551305 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.551345 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.551508 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.551665 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.551689 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.551708 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.551798 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.551837 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.654554 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.654630 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.654685 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qt76\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-kube-api-access-8qt76\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.654726 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.654754 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.654999 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.655057 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.655210 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.655464 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.655503 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.655531 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.655621 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.655652 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.655734 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.656033 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.656090 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.660272 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.660560 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.660722 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.661391 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.661520 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.661867 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.663009 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.663484 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.664471 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.664910 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.665404 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.665383 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.666423 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.667591 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.670125 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.672531 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qt76\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-kube-api-access-8qt76\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lncts\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:16 crc kubenswrapper[4824]: I1209 10:28:16.857518 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:28:17 crc kubenswrapper[4824]: I1209 10:28:17.380447 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts"] Dec 09 10:28:17 crc kubenswrapper[4824]: I1209 10:28:17.390891 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 10:28:17 crc kubenswrapper[4824]: I1209 10:28:17.406118 4824 scope.go:117] "RemoveContainer" containerID="d0ab6558e5bc53fd8b72cf22f8e4de08649e93342f3afbc2f9b46f91abd831ba" Dec 09 10:28:17 crc kubenswrapper[4824]: I1209 10:28:17.416426 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" event={"ID":"332799e5-f463-48ab-be97-77329e48b07d","Type":"ContainerStarted","Data":"54c671b529d21c55b128afd0bbdf571e636ddfd30709962c21c94216547af2e0"} Dec 09 10:28:17 crc kubenswrapper[4824]: I1209 10:28:17.457527 4824 scope.go:117] "RemoveContainer" containerID="a9c3600fe3735139a42bd8d826c49d7371117ce269bbb7d36569e4497ca5a539" Dec 09 10:28:18 crc kubenswrapper[4824]: I1209 10:28:18.429598 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" event={"ID":"332799e5-f463-48ab-be97-77329e48b07d","Type":"ContainerStarted","Data":"ea8005863281d5a97c17ec2d2778b94599e468a06a4f13d726c55e397437e6df"} Dec 09 10:28:18 crc kubenswrapper[4824]: I1209 10:28:18.468173 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" podStartSLOduration=2.013765144 podStartE2EDuration="2.468147824s" podCreationTimestamp="2025-12-09 10:28:16 +0000 UTC" firstStartedPulling="2025-12-09 10:28:17.390633164 +0000 UTC m=+2453.725137831" lastFinishedPulling="2025-12-09 10:28:17.845015844 +0000 UTC m=+2454.179520511" observedRunningTime="2025-12-09 10:28:18.451685709 +0000 UTC m=+2454.786190376" watchObservedRunningTime="2025-12-09 10:28:18.468147824 +0000 UTC m=+2454.802652491" Dec 09 10:28:24 crc kubenswrapper[4824]: I1209 10:28:24.910629 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:28:24 crc kubenswrapper[4824]: E1209 10:28:24.911438 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:28:39 crc kubenswrapper[4824]: I1209 10:28:39.911002 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:28:39 crc kubenswrapper[4824]: E1209 10:28:39.911991 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:28:54 crc kubenswrapper[4824]: I1209 10:28:54.911028 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:28:54 crc kubenswrapper[4824]: E1209 10:28:54.911993 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:29:06 crc kubenswrapper[4824]: I1209 10:29:06.511964 4824 generic.go:334] "Generic (PLEG): container finished" podID="332799e5-f463-48ab-be97-77329e48b07d" containerID="ea8005863281d5a97c17ec2d2778b94599e468a06a4f13d726c55e397437e6df" exitCode=0 Dec 09 10:29:06 crc kubenswrapper[4824]: I1209 10:29:06.512073 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" event={"ID":"332799e5-f463-48ab-be97-77329e48b07d","Type":"ContainerDied","Data":"ea8005863281d5a97c17ec2d2778b94599e468a06a4f13d726c55e397437e6df"} Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.051490 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.146488 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-bootstrap-combined-ca-bundle\") pod \"332799e5-f463-48ab-be97-77329e48b07d\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.146647 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-nova-combined-ca-bundle\") pod \"332799e5-f463-48ab-be97-77329e48b07d\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.146733 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-ssh-key\") pod \"332799e5-f463-48ab-be97-77329e48b07d\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.146793 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"332799e5-f463-48ab-be97-77329e48b07d\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.146839 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qt76\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-kube-api-access-8qt76\") pod \"332799e5-f463-48ab-be97-77329e48b07d\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.146867 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-ovn-combined-ca-bundle\") pod \"332799e5-f463-48ab-be97-77329e48b07d\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.146944 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-telemetry-power-monitoring-combined-ca-bundle\") pod \"332799e5-f463-48ab-be97-77329e48b07d\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.147044 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"332799e5-f463-48ab-be97-77329e48b07d\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.147069 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"332799e5-f463-48ab-be97-77329e48b07d\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.147099 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-telemetry-combined-ca-bundle\") pod \"332799e5-f463-48ab-be97-77329e48b07d\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.147146 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-inventory\") pod \"332799e5-f463-48ab-be97-77329e48b07d\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.147180 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-ovn-default-certs-0\") pod \"332799e5-f463-48ab-be97-77329e48b07d\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.147220 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-repo-setup-combined-ca-bundle\") pod \"332799e5-f463-48ab-be97-77329e48b07d\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.147266 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-neutron-metadata-combined-ca-bundle\") pod \"332799e5-f463-48ab-be97-77329e48b07d\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.147301 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"332799e5-f463-48ab-be97-77329e48b07d\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.147330 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-libvirt-combined-ca-bundle\") pod \"332799e5-f463-48ab-be97-77329e48b07d\" (UID: \"332799e5-f463-48ab-be97-77329e48b07d\") " Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.155746 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0") pod "332799e5-f463-48ab-be97-77329e48b07d" (UID: "332799e5-f463-48ab-be97-77329e48b07d"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.155955 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "332799e5-f463-48ab-be97-77329e48b07d" (UID: "332799e5-f463-48ab-be97-77329e48b07d"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.160725 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "332799e5-f463-48ab-be97-77329e48b07d" (UID: "332799e5-f463-48ab-be97-77329e48b07d"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.163006 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "332799e5-f463-48ab-be97-77329e48b07d" (UID: "332799e5-f463-48ab-be97-77329e48b07d"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.163054 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "332799e5-f463-48ab-be97-77329e48b07d" (UID: "332799e5-f463-48ab-be97-77329e48b07d"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.163199 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "332799e5-f463-48ab-be97-77329e48b07d" (UID: "332799e5-f463-48ab-be97-77329e48b07d"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.163193 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "332799e5-f463-48ab-be97-77329e48b07d" (UID: "332799e5-f463-48ab-be97-77329e48b07d"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.163240 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-kube-api-access-8qt76" (OuterVolumeSpecName: "kube-api-access-8qt76") pod "332799e5-f463-48ab-be97-77329e48b07d" (UID: "332799e5-f463-48ab-be97-77329e48b07d"). InnerVolumeSpecName "kube-api-access-8qt76". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.163232 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "332799e5-f463-48ab-be97-77329e48b07d" (UID: "332799e5-f463-48ab-be97-77329e48b07d"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.163365 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "332799e5-f463-48ab-be97-77329e48b07d" (UID: "332799e5-f463-48ab-be97-77329e48b07d"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.163367 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "332799e5-f463-48ab-be97-77329e48b07d" (UID: "332799e5-f463-48ab-be97-77329e48b07d"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.163577 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "332799e5-f463-48ab-be97-77329e48b07d" (UID: "332799e5-f463-48ab-be97-77329e48b07d"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.164238 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "332799e5-f463-48ab-be97-77329e48b07d" (UID: "332799e5-f463-48ab-be97-77329e48b07d"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.165377 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "332799e5-f463-48ab-be97-77329e48b07d" (UID: "332799e5-f463-48ab-be97-77329e48b07d"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.193459 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-inventory" (OuterVolumeSpecName: "inventory") pod "332799e5-f463-48ab-be97-77329e48b07d" (UID: "332799e5-f463-48ab-be97-77329e48b07d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.194458 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "332799e5-f463-48ab-be97-77329e48b07d" (UID: "332799e5-f463-48ab-be97-77329e48b07d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.252764 4824 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.252817 4824 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.252829 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.252839 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.252851 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qt76\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-kube-api-access-8qt76\") on node \"crc\" DevicePath \"\"" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.252864 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.252877 4824 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.252891 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.252904 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.252915 4824 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.252926 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.252936 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.252947 4824 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.252963 4824 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.252979 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/332799e5-f463-48ab-be97-77329e48b07d-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.252993 4824 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332799e5-f463-48ab-be97-77329e48b07d-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.536265 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" event={"ID":"332799e5-f463-48ab-be97-77329e48b07d","Type":"ContainerDied","Data":"54c671b529d21c55b128afd0bbdf571e636ddfd30709962c21c94216547af2e0"} Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.536325 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54c671b529d21c55b128afd0bbdf571e636ddfd30709962c21c94216547af2e0" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.536392 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lncts" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.901596 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh"] Dec 09 10:29:08 crc kubenswrapper[4824]: E1209 10:29:08.902142 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="332799e5-f463-48ab-be97-77329e48b07d" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.902171 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="332799e5-f463-48ab-be97-77329e48b07d" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.902448 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="332799e5-f463-48ab-be97-77329e48b07d" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.903311 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.909753 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.909879 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.910109 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.910124 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.910204 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5nzld" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.911372 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:29:08 crc kubenswrapper[4824]: E1209 10:29:08.911762 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.922752 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh"] Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.968169 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c209e08-d978-483b-88b8-2f0ef683d62e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dvksh\" (UID: \"5c209e08-d978-483b-88b8-2f0ef683d62e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.968526 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c209e08-d978-483b-88b8-2f0ef683d62e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dvksh\" (UID: \"5c209e08-d978-483b-88b8-2f0ef683d62e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.968560 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c209e08-d978-483b-88b8-2f0ef683d62e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dvksh\" (UID: \"5c209e08-d978-483b-88b8-2f0ef683d62e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.968587 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5c209e08-d978-483b-88b8-2f0ef683d62e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dvksh\" (UID: \"5c209e08-d978-483b-88b8-2f0ef683d62e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh" Dec 09 10:29:08 crc kubenswrapper[4824]: I1209 10:29:08.968636 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rd6gv\" (UniqueName: \"kubernetes.io/projected/5c209e08-d978-483b-88b8-2f0ef683d62e-kube-api-access-rd6gv\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dvksh\" (UID: \"5c209e08-d978-483b-88b8-2f0ef683d62e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh" Dec 09 10:29:09 crc kubenswrapper[4824]: I1209 10:29:09.070719 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c209e08-d978-483b-88b8-2f0ef683d62e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dvksh\" (UID: \"5c209e08-d978-483b-88b8-2f0ef683d62e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh" Dec 09 10:29:09 crc kubenswrapper[4824]: I1209 10:29:09.070877 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c209e08-d978-483b-88b8-2f0ef683d62e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dvksh\" (UID: \"5c209e08-d978-483b-88b8-2f0ef683d62e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh" Dec 09 10:29:09 crc kubenswrapper[4824]: I1209 10:29:09.070911 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c209e08-d978-483b-88b8-2f0ef683d62e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dvksh\" (UID: \"5c209e08-d978-483b-88b8-2f0ef683d62e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh" Dec 09 10:29:09 crc kubenswrapper[4824]: I1209 10:29:09.070939 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5c209e08-d978-483b-88b8-2f0ef683d62e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dvksh\" (UID: \"5c209e08-d978-483b-88b8-2f0ef683d62e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh" Dec 09 10:29:09 crc kubenswrapper[4824]: I1209 10:29:09.070975 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rd6gv\" (UniqueName: \"kubernetes.io/projected/5c209e08-d978-483b-88b8-2f0ef683d62e-kube-api-access-rd6gv\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dvksh\" (UID: \"5c209e08-d978-483b-88b8-2f0ef683d62e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh" Dec 09 10:29:09 crc kubenswrapper[4824]: I1209 10:29:09.072370 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5c209e08-d978-483b-88b8-2f0ef683d62e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dvksh\" (UID: \"5c209e08-d978-483b-88b8-2f0ef683d62e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh" Dec 09 10:29:09 crc kubenswrapper[4824]: I1209 10:29:09.077292 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c209e08-d978-483b-88b8-2f0ef683d62e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dvksh\" (UID: \"5c209e08-d978-483b-88b8-2f0ef683d62e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh" Dec 09 10:29:09 crc kubenswrapper[4824]: I1209 10:29:09.078644 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c209e08-d978-483b-88b8-2f0ef683d62e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dvksh\" (UID: \"5c209e08-d978-483b-88b8-2f0ef683d62e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh" Dec 09 10:29:09 crc kubenswrapper[4824]: I1209 10:29:09.083267 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c209e08-d978-483b-88b8-2f0ef683d62e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dvksh\" (UID: \"5c209e08-d978-483b-88b8-2f0ef683d62e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh" Dec 09 10:29:09 crc kubenswrapper[4824]: I1209 10:29:09.092093 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rd6gv\" (UniqueName: \"kubernetes.io/projected/5c209e08-d978-483b-88b8-2f0ef683d62e-kube-api-access-rd6gv\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dvksh\" (UID: \"5c209e08-d978-483b-88b8-2f0ef683d62e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh" Dec 09 10:29:09 crc kubenswrapper[4824]: I1209 10:29:09.225163 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh" Dec 09 10:29:09 crc kubenswrapper[4824]: I1209 10:29:09.808999 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh"] Dec 09 10:29:10 crc kubenswrapper[4824]: I1209 10:29:10.563477 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh" event={"ID":"5c209e08-d978-483b-88b8-2f0ef683d62e","Type":"ContainerStarted","Data":"01a734bc441edc2417175205a13060853085279d9f3353971b71afd2c6d3a735"} Dec 09 10:29:11 crc kubenswrapper[4824]: I1209 10:29:11.576245 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh" event={"ID":"5c209e08-d978-483b-88b8-2f0ef683d62e","Type":"ContainerStarted","Data":"733c1cfe764f6b4c0f5881cb31fd6c566319babe4614f8bc9f59cbebb8b7f206"} Dec 09 10:29:11 crc kubenswrapper[4824]: I1209 10:29:11.595411 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh" podStartSLOduration=3.124904256 podStartE2EDuration="3.59539188s" podCreationTimestamp="2025-12-09 10:29:08 +0000 UTC" firstStartedPulling="2025-12-09 10:29:09.812110183 +0000 UTC m=+2506.146614850" lastFinishedPulling="2025-12-09 10:29:10.282597807 +0000 UTC m=+2506.617102474" observedRunningTime="2025-12-09 10:29:11.592744427 +0000 UTC m=+2507.927249104" watchObservedRunningTime="2025-12-09 10:29:11.59539188 +0000 UTC m=+2507.929896547" Dec 09 10:29:21 crc kubenswrapper[4824]: I1209 10:29:21.911822 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:29:21 crc kubenswrapper[4824]: E1209 10:29:21.912795 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:29:29 crc kubenswrapper[4824]: I1209 10:29:29.742302 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sg8wk"] Dec 09 10:29:29 crc kubenswrapper[4824]: I1209 10:29:29.745979 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sg8wk" Dec 09 10:29:29 crc kubenswrapper[4824]: I1209 10:29:29.761689 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sg8wk"] Dec 09 10:29:29 crc kubenswrapper[4824]: I1209 10:29:29.825765 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fb78d54-47e2-415e-b8f4-6e5d88530082-catalog-content\") pod \"redhat-marketplace-sg8wk\" (UID: \"5fb78d54-47e2-415e-b8f4-6e5d88530082\") " pod="openshift-marketplace/redhat-marketplace-sg8wk" Dec 09 10:29:29 crc kubenswrapper[4824]: I1209 10:29:29.825997 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fb78d54-47e2-415e-b8f4-6e5d88530082-utilities\") pod \"redhat-marketplace-sg8wk\" (UID: \"5fb78d54-47e2-415e-b8f4-6e5d88530082\") " pod="openshift-marketplace/redhat-marketplace-sg8wk" Dec 09 10:29:29 crc kubenswrapper[4824]: I1209 10:29:29.826281 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jq7b\" (UniqueName: \"kubernetes.io/projected/5fb78d54-47e2-415e-b8f4-6e5d88530082-kube-api-access-4jq7b\") pod \"redhat-marketplace-sg8wk\" (UID: \"5fb78d54-47e2-415e-b8f4-6e5d88530082\") " pod="openshift-marketplace/redhat-marketplace-sg8wk" Dec 09 10:29:29 crc kubenswrapper[4824]: I1209 10:29:29.928712 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jq7b\" (UniqueName: \"kubernetes.io/projected/5fb78d54-47e2-415e-b8f4-6e5d88530082-kube-api-access-4jq7b\") pod \"redhat-marketplace-sg8wk\" (UID: \"5fb78d54-47e2-415e-b8f4-6e5d88530082\") " pod="openshift-marketplace/redhat-marketplace-sg8wk" Dec 09 10:29:29 crc kubenswrapper[4824]: I1209 10:29:29.928967 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fb78d54-47e2-415e-b8f4-6e5d88530082-catalog-content\") pod \"redhat-marketplace-sg8wk\" (UID: \"5fb78d54-47e2-415e-b8f4-6e5d88530082\") " pod="openshift-marketplace/redhat-marketplace-sg8wk" Dec 09 10:29:29 crc kubenswrapper[4824]: I1209 10:29:29.929039 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fb78d54-47e2-415e-b8f4-6e5d88530082-utilities\") pod \"redhat-marketplace-sg8wk\" (UID: \"5fb78d54-47e2-415e-b8f4-6e5d88530082\") " pod="openshift-marketplace/redhat-marketplace-sg8wk" Dec 09 10:29:29 crc kubenswrapper[4824]: I1209 10:29:29.929573 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fb78d54-47e2-415e-b8f4-6e5d88530082-utilities\") pod \"redhat-marketplace-sg8wk\" (UID: \"5fb78d54-47e2-415e-b8f4-6e5d88530082\") " pod="openshift-marketplace/redhat-marketplace-sg8wk" Dec 09 10:29:29 crc kubenswrapper[4824]: I1209 10:29:29.929839 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fb78d54-47e2-415e-b8f4-6e5d88530082-catalog-content\") pod \"redhat-marketplace-sg8wk\" (UID: \"5fb78d54-47e2-415e-b8f4-6e5d88530082\") " pod="openshift-marketplace/redhat-marketplace-sg8wk" Dec 09 10:29:29 crc kubenswrapper[4824]: I1209 10:29:29.948162 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jq7b\" (UniqueName: \"kubernetes.io/projected/5fb78d54-47e2-415e-b8f4-6e5d88530082-kube-api-access-4jq7b\") pod \"redhat-marketplace-sg8wk\" (UID: \"5fb78d54-47e2-415e-b8f4-6e5d88530082\") " pod="openshift-marketplace/redhat-marketplace-sg8wk" Dec 09 10:29:30 crc kubenswrapper[4824]: I1209 10:29:30.073957 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sg8wk" Dec 09 10:29:30 crc kubenswrapper[4824]: I1209 10:29:30.587234 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sg8wk"] Dec 09 10:29:30 crc kubenswrapper[4824]: W1209 10:29:30.593617 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fb78d54_47e2_415e_b8f4_6e5d88530082.slice/crio-d3f5146ff747bc259e7a91fe77933ec40ba1cbb35b6f569ae1e6776f63ef9e25 WatchSource:0}: Error finding container d3f5146ff747bc259e7a91fe77933ec40ba1cbb35b6f569ae1e6776f63ef9e25: Status 404 returned error can't find the container with id d3f5146ff747bc259e7a91fe77933ec40ba1cbb35b6f569ae1e6776f63ef9e25 Dec 09 10:29:30 crc kubenswrapper[4824]: I1209 10:29:30.970792 4824 generic.go:334] "Generic (PLEG): container finished" podID="5fb78d54-47e2-415e-b8f4-6e5d88530082" containerID="141813e37b1e74173144b044426b84cf7a7b39ff457c9c9a6acb551e7ae5556e" exitCode=0 Dec 09 10:29:30 crc kubenswrapper[4824]: I1209 10:29:30.970873 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sg8wk" event={"ID":"5fb78d54-47e2-415e-b8f4-6e5d88530082","Type":"ContainerDied","Data":"141813e37b1e74173144b044426b84cf7a7b39ff457c9c9a6acb551e7ae5556e"} Dec 09 10:29:30 crc kubenswrapper[4824]: I1209 10:29:30.971203 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sg8wk" event={"ID":"5fb78d54-47e2-415e-b8f4-6e5d88530082","Type":"ContainerStarted","Data":"d3f5146ff747bc259e7a91fe77933ec40ba1cbb35b6f569ae1e6776f63ef9e25"} Dec 09 10:29:32 crc kubenswrapper[4824]: I1209 10:29:32.993287 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sg8wk" event={"ID":"5fb78d54-47e2-415e-b8f4-6e5d88530082","Type":"ContainerStarted","Data":"c03f03f133836c3e3e1d6241e698895b560c3326844cae1f022547cd6864fb8d"} Dec 09 10:29:35 crc kubenswrapper[4824]: I1209 10:29:35.015845 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sg8wk" event={"ID":"5fb78d54-47e2-415e-b8f4-6e5d88530082","Type":"ContainerDied","Data":"c03f03f133836c3e3e1d6241e698895b560c3326844cae1f022547cd6864fb8d"} Dec 09 10:29:35 crc kubenswrapper[4824]: I1209 10:29:35.015769 4824 generic.go:334] "Generic (PLEG): container finished" podID="5fb78d54-47e2-415e-b8f4-6e5d88530082" containerID="c03f03f133836c3e3e1d6241e698895b560c3326844cae1f022547cd6864fb8d" exitCode=0 Dec 09 10:29:36 crc kubenswrapper[4824]: I1209 10:29:36.911022 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:29:36 crc kubenswrapper[4824]: E1209 10:29:36.911864 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:29:37 crc kubenswrapper[4824]: I1209 10:29:37.045674 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sg8wk" event={"ID":"5fb78d54-47e2-415e-b8f4-6e5d88530082","Type":"ContainerStarted","Data":"f48986e91e288c42f4e017b823d4892fba9cddafc7090f2bf3bdbd6d7862ece0"} Dec 09 10:29:37 crc kubenswrapper[4824]: I1209 10:29:37.073231 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sg8wk" podStartSLOduration=3.271390119 podStartE2EDuration="8.073205679s" podCreationTimestamp="2025-12-09 10:29:29 +0000 UTC" firstStartedPulling="2025-12-09 10:29:30.973313833 +0000 UTC m=+2527.307818490" lastFinishedPulling="2025-12-09 10:29:35.775129383 +0000 UTC m=+2532.109634050" observedRunningTime="2025-12-09 10:29:37.065835399 +0000 UTC m=+2533.400340076" watchObservedRunningTime="2025-12-09 10:29:37.073205679 +0000 UTC m=+2533.407710356" Dec 09 10:29:40 crc kubenswrapper[4824]: I1209 10:29:40.075741 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sg8wk" Dec 09 10:29:40 crc kubenswrapper[4824]: I1209 10:29:40.076348 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sg8wk" Dec 09 10:29:40 crc kubenswrapper[4824]: I1209 10:29:40.151060 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sg8wk" Dec 09 10:29:41 crc kubenswrapper[4824]: I1209 10:29:41.142016 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sg8wk" Dec 09 10:29:41 crc kubenswrapper[4824]: I1209 10:29:41.197757 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sg8wk"] Dec 09 10:29:43 crc kubenswrapper[4824]: I1209 10:29:43.110326 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sg8wk" podUID="5fb78d54-47e2-415e-b8f4-6e5d88530082" containerName="registry-server" containerID="cri-o://f48986e91e288c42f4e017b823d4892fba9cddafc7090f2bf3bdbd6d7862ece0" gracePeriod=2 Dec 09 10:29:43 crc kubenswrapper[4824]: I1209 10:29:43.875851 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sg8wk" Dec 09 10:29:43 crc kubenswrapper[4824]: I1209 10:29:43.992716 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jq7b\" (UniqueName: \"kubernetes.io/projected/5fb78d54-47e2-415e-b8f4-6e5d88530082-kube-api-access-4jq7b\") pod \"5fb78d54-47e2-415e-b8f4-6e5d88530082\" (UID: \"5fb78d54-47e2-415e-b8f4-6e5d88530082\") " Dec 09 10:29:43 crc kubenswrapper[4824]: I1209 10:29:43.993074 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fb78d54-47e2-415e-b8f4-6e5d88530082-utilities\") pod \"5fb78d54-47e2-415e-b8f4-6e5d88530082\" (UID: \"5fb78d54-47e2-415e-b8f4-6e5d88530082\") " Dec 09 10:29:43 crc kubenswrapper[4824]: I1209 10:29:43.993198 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fb78d54-47e2-415e-b8f4-6e5d88530082-catalog-content\") pod \"5fb78d54-47e2-415e-b8f4-6e5d88530082\" (UID: \"5fb78d54-47e2-415e-b8f4-6e5d88530082\") " Dec 09 10:29:43 crc kubenswrapper[4824]: I1209 10:29:43.994075 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fb78d54-47e2-415e-b8f4-6e5d88530082-utilities" (OuterVolumeSpecName: "utilities") pod "5fb78d54-47e2-415e-b8f4-6e5d88530082" (UID: "5fb78d54-47e2-415e-b8f4-6e5d88530082"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:29:44 crc kubenswrapper[4824]: I1209 10:29:44.000045 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fb78d54-47e2-415e-b8f4-6e5d88530082-kube-api-access-4jq7b" (OuterVolumeSpecName: "kube-api-access-4jq7b") pod "5fb78d54-47e2-415e-b8f4-6e5d88530082" (UID: "5fb78d54-47e2-415e-b8f4-6e5d88530082"). InnerVolumeSpecName "kube-api-access-4jq7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:29:44 crc kubenswrapper[4824]: I1209 10:29:44.015333 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fb78d54-47e2-415e-b8f4-6e5d88530082-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5fb78d54-47e2-415e-b8f4-6e5d88530082" (UID: "5fb78d54-47e2-415e-b8f4-6e5d88530082"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:29:44 crc kubenswrapper[4824]: I1209 10:29:44.096112 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fb78d54-47e2-415e-b8f4-6e5d88530082-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 10:29:44 crc kubenswrapper[4824]: I1209 10:29:44.096151 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fb78d54-47e2-415e-b8f4-6e5d88530082-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 10:29:44 crc kubenswrapper[4824]: I1209 10:29:44.096165 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jq7b\" (UniqueName: \"kubernetes.io/projected/5fb78d54-47e2-415e-b8f4-6e5d88530082-kube-api-access-4jq7b\") on node \"crc\" DevicePath \"\"" Dec 09 10:29:44 crc kubenswrapper[4824]: I1209 10:29:44.123708 4824 generic.go:334] "Generic (PLEG): container finished" podID="5fb78d54-47e2-415e-b8f4-6e5d88530082" containerID="f48986e91e288c42f4e017b823d4892fba9cddafc7090f2bf3bdbd6d7862ece0" exitCode=0 Dec 09 10:29:44 crc kubenswrapper[4824]: I1209 10:29:44.123746 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sg8wk" event={"ID":"5fb78d54-47e2-415e-b8f4-6e5d88530082","Type":"ContainerDied","Data":"f48986e91e288c42f4e017b823d4892fba9cddafc7090f2bf3bdbd6d7862ece0"} Dec 09 10:29:44 crc kubenswrapper[4824]: I1209 10:29:44.123768 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sg8wk" Dec 09 10:29:44 crc kubenswrapper[4824]: I1209 10:29:44.123812 4824 scope.go:117] "RemoveContainer" containerID="f48986e91e288c42f4e017b823d4892fba9cddafc7090f2bf3bdbd6d7862ece0" Dec 09 10:29:44 crc kubenswrapper[4824]: I1209 10:29:44.123771 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sg8wk" event={"ID":"5fb78d54-47e2-415e-b8f4-6e5d88530082","Type":"ContainerDied","Data":"d3f5146ff747bc259e7a91fe77933ec40ba1cbb35b6f569ae1e6776f63ef9e25"} Dec 09 10:29:44 crc kubenswrapper[4824]: I1209 10:29:44.153044 4824 scope.go:117] "RemoveContainer" containerID="c03f03f133836c3e3e1d6241e698895b560c3326844cae1f022547cd6864fb8d" Dec 09 10:29:44 crc kubenswrapper[4824]: I1209 10:29:44.162162 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sg8wk"] Dec 09 10:29:44 crc kubenswrapper[4824]: I1209 10:29:44.174927 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sg8wk"] Dec 09 10:29:44 crc kubenswrapper[4824]: I1209 10:29:44.191209 4824 scope.go:117] "RemoveContainer" containerID="141813e37b1e74173144b044426b84cf7a7b39ff457c9c9a6acb551e7ae5556e" Dec 09 10:29:44 crc kubenswrapper[4824]: I1209 10:29:44.233751 4824 scope.go:117] "RemoveContainer" containerID="f48986e91e288c42f4e017b823d4892fba9cddafc7090f2bf3bdbd6d7862ece0" Dec 09 10:29:44 crc kubenswrapper[4824]: E1209 10:29:44.234215 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f48986e91e288c42f4e017b823d4892fba9cddafc7090f2bf3bdbd6d7862ece0\": container with ID starting with f48986e91e288c42f4e017b823d4892fba9cddafc7090f2bf3bdbd6d7862ece0 not found: ID does not exist" containerID="f48986e91e288c42f4e017b823d4892fba9cddafc7090f2bf3bdbd6d7862ece0" Dec 09 10:29:44 crc kubenswrapper[4824]: I1209 10:29:44.234247 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f48986e91e288c42f4e017b823d4892fba9cddafc7090f2bf3bdbd6d7862ece0"} err="failed to get container status \"f48986e91e288c42f4e017b823d4892fba9cddafc7090f2bf3bdbd6d7862ece0\": rpc error: code = NotFound desc = could not find container \"f48986e91e288c42f4e017b823d4892fba9cddafc7090f2bf3bdbd6d7862ece0\": container with ID starting with f48986e91e288c42f4e017b823d4892fba9cddafc7090f2bf3bdbd6d7862ece0 not found: ID does not exist" Dec 09 10:29:44 crc kubenswrapper[4824]: I1209 10:29:44.234285 4824 scope.go:117] "RemoveContainer" containerID="c03f03f133836c3e3e1d6241e698895b560c3326844cae1f022547cd6864fb8d" Dec 09 10:29:44 crc kubenswrapper[4824]: E1209 10:29:44.234612 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c03f03f133836c3e3e1d6241e698895b560c3326844cae1f022547cd6864fb8d\": container with ID starting with c03f03f133836c3e3e1d6241e698895b560c3326844cae1f022547cd6864fb8d not found: ID does not exist" containerID="c03f03f133836c3e3e1d6241e698895b560c3326844cae1f022547cd6864fb8d" Dec 09 10:29:44 crc kubenswrapper[4824]: I1209 10:29:44.234685 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c03f03f133836c3e3e1d6241e698895b560c3326844cae1f022547cd6864fb8d"} err="failed to get container status \"c03f03f133836c3e3e1d6241e698895b560c3326844cae1f022547cd6864fb8d\": rpc error: code = NotFound desc = could not find container \"c03f03f133836c3e3e1d6241e698895b560c3326844cae1f022547cd6864fb8d\": container with ID starting with c03f03f133836c3e3e1d6241e698895b560c3326844cae1f022547cd6864fb8d not found: ID does not exist" Dec 09 10:29:44 crc kubenswrapper[4824]: I1209 10:29:44.234702 4824 scope.go:117] "RemoveContainer" containerID="141813e37b1e74173144b044426b84cf7a7b39ff457c9c9a6acb551e7ae5556e" Dec 09 10:29:44 crc kubenswrapper[4824]: E1209 10:29:44.235318 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"141813e37b1e74173144b044426b84cf7a7b39ff457c9c9a6acb551e7ae5556e\": container with ID starting with 141813e37b1e74173144b044426b84cf7a7b39ff457c9c9a6acb551e7ae5556e not found: ID does not exist" containerID="141813e37b1e74173144b044426b84cf7a7b39ff457c9c9a6acb551e7ae5556e" Dec 09 10:29:44 crc kubenswrapper[4824]: I1209 10:29:44.235348 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"141813e37b1e74173144b044426b84cf7a7b39ff457c9c9a6acb551e7ae5556e"} err="failed to get container status \"141813e37b1e74173144b044426b84cf7a7b39ff457c9c9a6acb551e7ae5556e\": rpc error: code = NotFound desc = could not find container \"141813e37b1e74173144b044426b84cf7a7b39ff457c9c9a6acb551e7ae5556e\": container with ID starting with 141813e37b1e74173144b044426b84cf7a7b39ff457c9c9a6acb551e7ae5556e not found: ID does not exist" Dec 09 10:29:45 crc kubenswrapper[4824]: I1209 10:29:45.926983 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fb78d54-47e2-415e-b8f4-6e5d88530082" path="/var/lib/kubelet/pods/5fb78d54-47e2-415e-b8f4-6e5d88530082/volumes" Dec 09 10:29:47 crc kubenswrapper[4824]: I1209 10:29:47.911584 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:29:47 crc kubenswrapper[4824]: E1209 10:29:47.912372 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:29:58 crc kubenswrapper[4824]: I1209 10:29:58.699286 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8tjcr"] Dec 09 10:29:58 crc kubenswrapper[4824]: E1209 10:29:58.701717 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fb78d54-47e2-415e-b8f4-6e5d88530082" containerName="extract-utilities" Dec 09 10:29:58 crc kubenswrapper[4824]: I1209 10:29:58.701896 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fb78d54-47e2-415e-b8f4-6e5d88530082" containerName="extract-utilities" Dec 09 10:29:58 crc kubenswrapper[4824]: E1209 10:29:58.701997 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fb78d54-47e2-415e-b8f4-6e5d88530082" containerName="extract-content" Dec 09 10:29:58 crc kubenswrapper[4824]: I1209 10:29:58.702094 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fb78d54-47e2-415e-b8f4-6e5d88530082" containerName="extract-content" Dec 09 10:29:58 crc kubenswrapper[4824]: E1209 10:29:58.702183 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fb78d54-47e2-415e-b8f4-6e5d88530082" containerName="registry-server" Dec 09 10:29:58 crc kubenswrapper[4824]: I1209 10:29:58.702252 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fb78d54-47e2-415e-b8f4-6e5d88530082" containerName="registry-server" Dec 09 10:29:58 crc kubenswrapper[4824]: I1209 10:29:58.702727 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fb78d54-47e2-415e-b8f4-6e5d88530082" containerName="registry-server" Dec 09 10:29:58 crc kubenswrapper[4824]: I1209 10:29:58.705105 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8tjcr" Dec 09 10:29:58 crc kubenswrapper[4824]: I1209 10:29:58.712682 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8tjcr"] Dec 09 10:29:58 crc kubenswrapper[4824]: I1209 10:29:58.783030 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/def49a3f-d880-44de-a0dc-3b57abd944a7-catalog-content\") pod \"certified-operators-8tjcr\" (UID: \"def49a3f-d880-44de-a0dc-3b57abd944a7\") " pod="openshift-marketplace/certified-operators-8tjcr" Dec 09 10:29:58 crc kubenswrapper[4824]: I1209 10:29:58.784076 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbhzz\" (UniqueName: \"kubernetes.io/projected/def49a3f-d880-44de-a0dc-3b57abd944a7-kube-api-access-qbhzz\") pod \"certified-operators-8tjcr\" (UID: \"def49a3f-d880-44de-a0dc-3b57abd944a7\") " pod="openshift-marketplace/certified-operators-8tjcr" Dec 09 10:29:58 crc kubenswrapper[4824]: I1209 10:29:58.784809 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/def49a3f-d880-44de-a0dc-3b57abd944a7-utilities\") pod \"certified-operators-8tjcr\" (UID: \"def49a3f-d880-44de-a0dc-3b57abd944a7\") " pod="openshift-marketplace/certified-operators-8tjcr" Dec 09 10:29:58 crc kubenswrapper[4824]: I1209 10:29:58.887696 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/def49a3f-d880-44de-a0dc-3b57abd944a7-utilities\") pod \"certified-operators-8tjcr\" (UID: \"def49a3f-d880-44de-a0dc-3b57abd944a7\") " pod="openshift-marketplace/certified-operators-8tjcr" Dec 09 10:29:58 crc kubenswrapper[4824]: I1209 10:29:58.888031 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/def49a3f-d880-44de-a0dc-3b57abd944a7-catalog-content\") pod \"certified-operators-8tjcr\" (UID: \"def49a3f-d880-44de-a0dc-3b57abd944a7\") " pod="openshift-marketplace/certified-operators-8tjcr" Dec 09 10:29:58 crc kubenswrapper[4824]: I1209 10:29:58.888265 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbhzz\" (UniqueName: \"kubernetes.io/projected/def49a3f-d880-44de-a0dc-3b57abd944a7-kube-api-access-qbhzz\") pod \"certified-operators-8tjcr\" (UID: \"def49a3f-d880-44de-a0dc-3b57abd944a7\") " pod="openshift-marketplace/certified-operators-8tjcr" Dec 09 10:29:58 crc kubenswrapper[4824]: I1209 10:29:58.888381 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/def49a3f-d880-44de-a0dc-3b57abd944a7-utilities\") pod \"certified-operators-8tjcr\" (UID: \"def49a3f-d880-44de-a0dc-3b57abd944a7\") " pod="openshift-marketplace/certified-operators-8tjcr" Dec 09 10:29:58 crc kubenswrapper[4824]: I1209 10:29:58.888637 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/def49a3f-d880-44de-a0dc-3b57abd944a7-catalog-content\") pod \"certified-operators-8tjcr\" (UID: \"def49a3f-d880-44de-a0dc-3b57abd944a7\") " pod="openshift-marketplace/certified-operators-8tjcr" Dec 09 10:29:58 crc kubenswrapper[4824]: I1209 10:29:58.912033 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:29:58 crc kubenswrapper[4824]: E1209 10:29:58.912615 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:29:58 crc kubenswrapper[4824]: I1209 10:29:58.919664 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbhzz\" (UniqueName: \"kubernetes.io/projected/def49a3f-d880-44de-a0dc-3b57abd944a7-kube-api-access-qbhzz\") pod \"certified-operators-8tjcr\" (UID: \"def49a3f-d880-44de-a0dc-3b57abd944a7\") " pod="openshift-marketplace/certified-operators-8tjcr" Dec 09 10:29:59 crc kubenswrapper[4824]: I1209 10:29:59.028852 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8tjcr" Dec 09 10:29:59 crc kubenswrapper[4824]: I1209 10:29:59.691135 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8tjcr"] Dec 09 10:30:00 crc kubenswrapper[4824]: I1209 10:30:00.168983 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421270-8wp4p"] Dec 09 10:30:00 crc kubenswrapper[4824]: I1209 10:30:00.171759 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421270-8wp4p" Dec 09 10:30:00 crc kubenswrapper[4824]: I1209 10:30:00.174640 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 10:30:00 crc kubenswrapper[4824]: I1209 10:30:00.179940 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 10:30:00 crc kubenswrapper[4824]: I1209 10:30:00.181886 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421270-8wp4p"] Dec 09 10:30:00 crc kubenswrapper[4824]: I1209 10:30:00.300279 4824 generic.go:334] "Generic (PLEG): container finished" podID="def49a3f-d880-44de-a0dc-3b57abd944a7" containerID="b29b4bc9b85dfaf70bc8d10f7b43e6ff58283eedeacf7b2e1e8b6c713b251120" exitCode=0 Dec 09 10:30:00 crc kubenswrapper[4824]: I1209 10:30:00.300333 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8tjcr" event={"ID":"def49a3f-d880-44de-a0dc-3b57abd944a7","Type":"ContainerDied","Data":"b29b4bc9b85dfaf70bc8d10f7b43e6ff58283eedeacf7b2e1e8b6c713b251120"} Dec 09 10:30:00 crc kubenswrapper[4824]: I1209 10:30:00.300367 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8tjcr" event={"ID":"def49a3f-d880-44de-a0dc-3b57abd944a7","Type":"ContainerStarted","Data":"a54218b9137ffc86f264e89bb19eaad11fbeac3d8ced82f7ccf727cc240c07ec"} Dec 09 10:30:00 crc kubenswrapper[4824]: I1209 10:30:00.329663 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pl4p9\" (UniqueName: \"kubernetes.io/projected/65d6395b-1cca-479b-986c-49b2142e7a30-kube-api-access-pl4p9\") pod \"collect-profiles-29421270-8wp4p\" (UID: \"65d6395b-1cca-479b-986c-49b2142e7a30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421270-8wp4p" Dec 09 10:30:00 crc kubenswrapper[4824]: I1209 10:30:00.329970 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/65d6395b-1cca-479b-986c-49b2142e7a30-config-volume\") pod \"collect-profiles-29421270-8wp4p\" (UID: \"65d6395b-1cca-479b-986c-49b2142e7a30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421270-8wp4p" Dec 09 10:30:00 crc kubenswrapper[4824]: I1209 10:30:00.330330 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/65d6395b-1cca-479b-986c-49b2142e7a30-secret-volume\") pod \"collect-profiles-29421270-8wp4p\" (UID: \"65d6395b-1cca-479b-986c-49b2142e7a30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421270-8wp4p" Dec 09 10:30:00 crc kubenswrapper[4824]: I1209 10:30:00.433067 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/65d6395b-1cca-479b-986c-49b2142e7a30-secret-volume\") pod \"collect-profiles-29421270-8wp4p\" (UID: \"65d6395b-1cca-479b-986c-49b2142e7a30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421270-8wp4p" Dec 09 10:30:00 crc kubenswrapper[4824]: I1209 10:30:00.433326 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pl4p9\" (UniqueName: \"kubernetes.io/projected/65d6395b-1cca-479b-986c-49b2142e7a30-kube-api-access-pl4p9\") pod \"collect-profiles-29421270-8wp4p\" (UID: \"65d6395b-1cca-479b-986c-49b2142e7a30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421270-8wp4p" Dec 09 10:30:00 crc kubenswrapper[4824]: I1209 10:30:00.433438 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/65d6395b-1cca-479b-986c-49b2142e7a30-config-volume\") pod \"collect-profiles-29421270-8wp4p\" (UID: \"65d6395b-1cca-479b-986c-49b2142e7a30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421270-8wp4p" Dec 09 10:30:00 crc kubenswrapper[4824]: I1209 10:30:00.434369 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/65d6395b-1cca-479b-986c-49b2142e7a30-config-volume\") pod \"collect-profiles-29421270-8wp4p\" (UID: \"65d6395b-1cca-479b-986c-49b2142e7a30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421270-8wp4p" Dec 09 10:30:00 crc kubenswrapper[4824]: I1209 10:30:00.441270 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/65d6395b-1cca-479b-986c-49b2142e7a30-secret-volume\") pod \"collect-profiles-29421270-8wp4p\" (UID: \"65d6395b-1cca-479b-986c-49b2142e7a30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421270-8wp4p" Dec 09 10:30:00 crc kubenswrapper[4824]: I1209 10:30:00.464511 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pl4p9\" (UniqueName: \"kubernetes.io/projected/65d6395b-1cca-479b-986c-49b2142e7a30-kube-api-access-pl4p9\") pod \"collect-profiles-29421270-8wp4p\" (UID: \"65d6395b-1cca-479b-986c-49b2142e7a30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421270-8wp4p" Dec 09 10:30:00 crc kubenswrapper[4824]: I1209 10:30:00.503035 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421270-8wp4p" Dec 09 10:30:00 crc kubenswrapper[4824]: I1209 10:30:00.997265 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421270-8wp4p"] Dec 09 10:30:01 crc kubenswrapper[4824]: W1209 10:30:01.003850 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65d6395b_1cca_479b_986c_49b2142e7a30.slice/crio-ee9655ed7be4428e9139d4f549fe585d76f518e921d63d16a0da0e3059b0f8a1 WatchSource:0}: Error finding container ee9655ed7be4428e9139d4f549fe585d76f518e921d63d16a0da0e3059b0f8a1: Status 404 returned error can't find the container with id ee9655ed7be4428e9139d4f549fe585d76f518e921d63d16a0da0e3059b0f8a1 Dec 09 10:30:01 crc kubenswrapper[4824]: I1209 10:30:01.315963 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421270-8wp4p" event={"ID":"65d6395b-1cca-479b-986c-49b2142e7a30","Type":"ContainerStarted","Data":"91cc1a2de21bfb84fc95fdf995d849a2432267276a144e63a9c420151952f83e"} Dec 09 10:30:01 crc kubenswrapper[4824]: I1209 10:30:01.316570 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421270-8wp4p" event={"ID":"65d6395b-1cca-479b-986c-49b2142e7a30","Type":"ContainerStarted","Data":"ee9655ed7be4428e9139d4f549fe585d76f518e921d63d16a0da0e3059b0f8a1"} Dec 09 10:30:01 crc kubenswrapper[4824]: I1209 10:30:01.348395 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29421270-8wp4p" podStartSLOduration=1.348369581 podStartE2EDuration="1.348369581s" podCreationTimestamp="2025-12-09 10:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:30:01.336301125 +0000 UTC m=+2557.670805822" watchObservedRunningTime="2025-12-09 10:30:01.348369581 +0000 UTC m=+2557.682874248" Dec 09 10:30:02 crc kubenswrapper[4824]: I1209 10:30:02.330453 4824 generic.go:334] "Generic (PLEG): container finished" podID="65d6395b-1cca-479b-986c-49b2142e7a30" containerID="91cc1a2de21bfb84fc95fdf995d849a2432267276a144e63a9c420151952f83e" exitCode=0 Dec 09 10:30:02 crc kubenswrapper[4824]: I1209 10:30:02.330528 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421270-8wp4p" event={"ID":"65d6395b-1cca-479b-986c-49b2142e7a30","Type":"ContainerDied","Data":"91cc1a2de21bfb84fc95fdf995d849a2432267276a144e63a9c420151952f83e"} Dec 09 10:30:02 crc kubenswrapper[4824]: I1209 10:30:02.333395 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8tjcr" event={"ID":"def49a3f-d880-44de-a0dc-3b57abd944a7","Type":"ContainerStarted","Data":"b597982b7287c0c1e731110020c7b308df20d8fd3d28990ff5b8dd4b862c1455"} Dec 09 10:30:03 crc kubenswrapper[4824]: I1209 10:30:03.350304 4824 generic.go:334] "Generic (PLEG): container finished" podID="def49a3f-d880-44de-a0dc-3b57abd944a7" containerID="b597982b7287c0c1e731110020c7b308df20d8fd3d28990ff5b8dd4b862c1455" exitCode=0 Dec 09 10:30:03 crc kubenswrapper[4824]: I1209 10:30:03.350642 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8tjcr" event={"ID":"def49a3f-d880-44de-a0dc-3b57abd944a7","Type":"ContainerDied","Data":"b597982b7287c0c1e731110020c7b308df20d8fd3d28990ff5b8dd4b862c1455"} Dec 09 10:30:03 crc kubenswrapper[4824]: I1209 10:30:03.793241 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421270-8wp4p" Dec 09 10:30:03 crc kubenswrapper[4824]: I1209 10:30:03.924844 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/65d6395b-1cca-479b-986c-49b2142e7a30-config-volume\") pod \"65d6395b-1cca-479b-986c-49b2142e7a30\" (UID: \"65d6395b-1cca-479b-986c-49b2142e7a30\") " Dec 09 10:30:03 crc kubenswrapper[4824]: I1209 10:30:03.925082 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pl4p9\" (UniqueName: \"kubernetes.io/projected/65d6395b-1cca-479b-986c-49b2142e7a30-kube-api-access-pl4p9\") pod \"65d6395b-1cca-479b-986c-49b2142e7a30\" (UID: \"65d6395b-1cca-479b-986c-49b2142e7a30\") " Dec 09 10:30:03 crc kubenswrapper[4824]: I1209 10:30:03.925111 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/65d6395b-1cca-479b-986c-49b2142e7a30-secret-volume\") pod \"65d6395b-1cca-479b-986c-49b2142e7a30\" (UID: \"65d6395b-1cca-479b-986c-49b2142e7a30\") " Dec 09 10:30:03 crc kubenswrapper[4824]: I1209 10:30:03.926135 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65d6395b-1cca-479b-986c-49b2142e7a30-config-volume" (OuterVolumeSpecName: "config-volume") pod "65d6395b-1cca-479b-986c-49b2142e7a30" (UID: "65d6395b-1cca-479b-986c-49b2142e7a30"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:30:03 crc kubenswrapper[4824]: I1209 10:30:03.934442 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65d6395b-1cca-479b-986c-49b2142e7a30-kube-api-access-pl4p9" (OuterVolumeSpecName: "kube-api-access-pl4p9") pod "65d6395b-1cca-479b-986c-49b2142e7a30" (UID: "65d6395b-1cca-479b-986c-49b2142e7a30"). InnerVolumeSpecName "kube-api-access-pl4p9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:30:03 crc kubenswrapper[4824]: I1209 10:30:03.936578 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65d6395b-1cca-479b-986c-49b2142e7a30-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "65d6395b-1cca-479b-986c-49b2142e7a30" (UID: "65d6395b-1cca-479b-986c-49b2142e7a30"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:30:04 crc kubenswrapper[4824]: I1209 10:30:04.029605 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/65d6395b-1cca-479b-986c-49b2142e7a30-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 10:30:04 crc kubenswrapper[4824]: I1209 10:30:04.029865 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pl4p9\" (UniqueName: \"kubernetes.io/projected/65d6395b-1cca-479b-986c-49b2142e7a30-kube-api-access-pl4p9\") on node \"crc\" DevicePath \"\"" Dec 09 10:30:04 crc kubenswrapper[4824]: I1209 10:30:04.029883 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/65d6395b-1cca-479b-986c-49b2142e7a30-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 10:30:04 crc kubenswrapper[4824]: I1209 10:30:04.365312 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8tjcr" event={"ID":"def49a3f-d880-44de-a0dc-3b57abd944a7","Type":"ContainerStarted","Data":"5f05bcc8c045e0e2bb84f94b0edf7d86325baf534918e33f4fab6f7433975fe4"} Dec 09 10:30:04 crc kubenswrapper[4824]: I1209 10:30:04.385368 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421270-8wp4p" event={"ID":"65d6395b-1cca-479b-986c-49b2142e7a30","Type":"ContainerDied","Data":"ee9655ed7be4428e9139d4f549fe585d76f518e921d63d16a0da0e3059b0f8a1"} Dec 09 10:30:04 crc kubenswrapper[4824]: I1209 10:30:04.385410 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee9655ed7be4428e9139d4f549fe585d76f518e921d63d16a0da0e3059b0f8a1" Dec 09 10:30:04 crc kubenswrapper[4824]: I1209 10:30:04.385472 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421270-8wp4p" Dec 09 10:30:04 crc kubenswrapper[4824]: I1209 10:30:04.401123 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8tjcr" podStartSLOduration=2.912095998 podStartE2EDuration="6.401099572s" podCreationTimestamp="2025-12-09 10:29:58 +0000 UTC" firstStartedPulling="2025-12-09 10:30:00.302469141 +0000 UTC m=+2556.636973808" lastFinishedPulling="2025-12-09 10:30:03.791472715 +0000 UTC m=+2560.125977382" observedRunningTime="2025-12-09 10:30:04.398338996 +0000 UTC m=+2560.732843663" watchObservedRunningTime="2025-12-09 10:30:04.401099572 +0000 UTC m=+2560.735604259" Dec 09 10:30:04 crc kubenswrapper[4824]: I1209 10:30:04.451650 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421225-96fgb"] Dec 09 10:30:04 crc kubenswrapper[4824]: I1209 10:30:04.470856 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421225-96fgb"] Dec 09 10:30:05 crc kubenswrapper[4824]: I1209 10:30:05.941960 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="055ead87-3e57-4053-a026-24566879437c" path="/var/lib/kubelet/pods/055ead87-3e57-4053-a026-24566879437c/volumes" Dec 09 10:30:09 crc kubenswrapper[4824]: I1209 10:30:09.030680 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8tjcr" Dec 09 10:30:09 crc kubenswrapper[4824]: I1209 10:30:09.031329 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8tjcr" Dec 09 10:30:09 crc kubenswrapper[4824]: I1209 10:30:09.103906 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8tjcr" Dec 09 10:30:09 crc kubenswrapper[4824]: I1209 10:30:09.509907 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8tjcr" Dec 09 10:30:09 crc kubenswrapper[4824]: I1209 10:30:09.561337 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8tjcr"] Dec 09 10:30:10 crc kubenswrapper[4824]: I1209 10:30:10.910694 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:30:10 crc kubenswrapper[4824]: E1209 10:30:10.911427 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:30:11 crc kubenswrapper[4824]: I1209 10:30:11.485333 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8tjcr" podUID="def49a3f-d880-44de-a0dc-3b57abd944a7" containerName="registry-server" containerID="cri-o://5f05bcc8c045e0e2bb84f94b0edf7d86325baf534918e33f4fab6f7433975fe4" gracePeriod=2 Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.032325 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8tjcr" Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.179590 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbhzz\" (UniqueName: \"kubernetes.io/projected/def49a3f-d880-44de-a0dc-3b57abd944a7-kube-api-access-qbhzz\") pod \"def49a3f-d880-44de-a0dc-3b57abd944a7\" (UID: \"def49a3f-d880-44de-a0dc-3b57abd944a7\") " Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.180223 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/def49a3f-d880-44de-a0dc-3b57abd944a7-utilities\") pod \"def49a3f-d880-44de-a0dc-3b57abd944a7\" (UID: \"def49a3f-d880-44de-a0dc-3b57abd944a7\") " Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.180282 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/def49a3f-d880-44de-a0dc-3b57abd944a7-catalog-content\") pod \"def49a3f-d880-44de-a0dc-3b57abd944a7\" (UID: \"def49a3f-d880-44de-a0dc-3b57abd944a7\") " Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.180914 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/def49a3f-d880-44de-a0dc-3b57abd944a7-utilities" (OuterVolumeSpecName: "utilities") pod "def49a3f-d880-44de-a0dc-3b57abd944a7" (UID: "def49a3f-d880-44de-a0dc-3b57abd944a7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.182999 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/def49a3f-d880-44de-a0dc-3b57abd944a7-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.192065 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/def49a3f-d880-44de-a0dc-3b57abd944a7-kube-api-access-qbhzz" (OuterVolumeSpecName: "kube-api-access-qbhzz") pod "def49a3f-d880-44de-a0dc-3b57abd944a7" (UID: "def49a3f-d880-44de-a0dc-3b57abd944a7"). InnerVolumeSpecName "kube-api-access-qbhzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.241969 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/def49a3f-d880-44de-a0dc-3b57abd944a7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "def49a3f-d880-44de-a0dc-3b57abd944a7" (UID: "def49a3f-d880-44de-a0dc-3b57abd944a7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.286027 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/def49a3f-d880-44de-a0dc-3b57abd944a7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.286065 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbhzz\" (UniqueName: \"kubernetes.io/projected/def49a3f-d880-44de-a0dc-3b57abd944a7-kube-api-access-qbhzz\") on node \"crc\" DevicePath \"\"" Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.500077 4824 generic.go:334] "Generic (PLEG): container finished" podID="def49a3f-d880-44de-a0dc-3b57abd944a7" containerID="5f05bcc8c045e0e2bb84f94b0edf7d86325baf534918e33f4fab6f7433975fe4" exitCode=0 Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.500130 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8tjcr" event={"ID":"def49a3f-d880-44de-a0dc-3b57abd944a7","Type":"ContainerDied","Data":"5f05bcc8c045e0e2bb84f94b0edf7d86325baf534918e33f4fab6f7433975fe4"} Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.500167 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8tjcr" event={"ID":"def49a3f-d880-44de-a0dc-3b57abd944a7","Type":"ContainerDied","Data":"a54218b9137ffc86f264e89bb19eaad11fbeac3d8ced82f7ccf727cc240c07ec"} Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.500189 4824 scope.go:117] "RemoveContainer" containerID="5f05bcc8c045e0e2bb84f94b0edf7d86325baf534918e33f4fab6f7433975fe4" Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.500213 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8tjcr" Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.523620 4824 scope.go:117] "RemoveContainer" containerID="b597982b7287c0c1e731110020c7b308df20d8fd3d28990ff5b8dd4b862c1455" Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.545089 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8tjcr"] Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.559689 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8tjcr"] Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.564998 4824 scope.go:117] "RemoveContainer" containerID="b29b4bc9b85dfaf70bc8d10f7b43e6ff58283eedeacf7b2e1e8b6c713b251120" Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.609766 4824 scope.go:117] "RemoveContainer" containerID="5f05bcc8c045e0e2bb84f94b0edf7d86325baf534918e33f4fab6f7433975fe4" Dec 09 10:30:12 crc kubenswrapper[4824]: E1209 10:30:12.610467 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f05bcc8c045e0e2bb84f94b0edf7d86325baf534918e33f4fab6f7433975fe4\": container with ID starting with 5f05bcc8c045e0e2bb84f94b0edf7d86325baf534918e33f4fab6f7433975fe4 not found: ID does not exist" containerID="5f05bcc8c045e0e2bb84f94b0edf7d86325baf534918e33f4fab6f7433975fe4" Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.610506 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f05bcc8c045e0e2bb84f94b0edf7d86325baf534918e33f4fab6f7433975fe4"} err="failed to get container status \"5f05bcc8c045e0e2bb84f94b0edf7d86325baf534918e33f4fab6f7433975fe4\": rpc error: code = NotFound desc = could not find container \"5f05bcc8c045e0e2bb84f94b0edf7d86325baf534918e33f4fab6f7433975fe4\": container with ID starting with 5f05bcc8c045e0e2bb84f94b0edf7d86325baf534918e33f4fab6f7433975fe4 not found: ID does not exist" Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.610539 4824 scope.go:117] "RemoveContainer" containerID="b597982b7287c0c1e731110020c7b308df20d8fd3d28990ff5b8dd4b862c1455" Dec 09 10:30:12 crc kubenswrapper[4824]: E1209 10:30:12.611147 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b597982b7287c0c1e731110020c7b308df20d8fd3d28990ff5b8dd4b862c1455\": container with ID starting with b597982b7287c0c1e731110020c7b308df20d8fd3d28990ff5b8dd4b862c1455 not found: ID does not exist" containerID="b597982b7287c0c1e731110020c7b308df20d8fd3d28990ff5b8dd4b862c1455" Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.611177 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b597982b7287c0c1e731110020c7b308df20d8fd3d28990ff5b8dd4b862c1455"} err="failed to get container status \"b597982b7287c0c1e731110020c7b308df20d8fd3d28990ff5b8dd4b862c1455\": rpc error: code = NotFound desc = could not find container \"b597982b7287c0c1e731110020c7b308df20d8fd3d28990ff5b8dd4b862c1455\": container with ID starting with b597982b7287c0c1e731110020c7b308df20d8fd3d28990ff5b8dd4b862c1455 not found: ID does not exist" Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.611196 4824 scope.go:117] "RemoveContainer" containerID="b29b4bc9b85dfaf70bc8d10f7b43e6ff58283eedeacf7b2e1e8b6c713b251120" Dec 09 10:30:12 crc kubenswrapper[4824]: E1209 10:30:12.611519 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b29b4bc9b85dfaf70bc8d10f7b43e6ff58283eedeacf7b2e1e8b6c713b251120\": container with ID starting with b29b4bc9b85dfaf70bc8d10f7b43e6ff58283eedeacf7b2e1e8b6c713b251120 not found: ID does not exist" containerID="b29b4bc9b85dfaf70bc8d10f7b43e6ff58283eedeacf7b2e1e8b6c713b251120" Dec 09 10:30:12 crc kubenswrapper[4824]: I1209 10:30:12.611541 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b29b4bc9b85dfaf70bc8d10f7b43e6ff58283eedeacf7b2e1e8b6c713b251120"} err="failed to get container status \"b29b4bc9b85dfaf70bc8d10f7b43e6ff58283eedeacf7b2e1e8b6c713b251120\": rpc error: code = NotFound desc = could not find container \"b29b4bc9b85dfaf70bc8d10f7b43e6ff58283eedeacf7b2e1e8b6c713b251120\": container with ID starting with b29b4bc9b85dfaf70bc8d10f7b43e6ff58283eedeacf7b2e1e8b6c713b251120 not found: ID does not exist" Dec 09 10:30:13 crc kubenswrapper[4824]: I1209 10:30:13.926901 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="def49a3f-d880-44de-a0dc-3b57abd944a7" path="/var/lib/kubelet/pods/def49a3f-d880-44de-a0dc-3b57abd944a7/volumes" Dec 09 10:30:17 crc kubenswrapper[4824]: I1209 10:30:17.592904 4824 scope.go:117] "RemoveContainer" containerID="91eeeddf7380a1f76b21a29c04f33b21a23c21087506d6559bdb21b8ad896cac" Dec 09 10:30:21 crc kubenswrapper[4824]: I1209 10:30:21.603363 4824 generic.go:334] "Generic (PLEG): container finished" podID="5c209e08-d978-483b-88b8-2f0ef683d62e" containerID="733c1cfe764f6b4c0f5881cb31fd6c566319babe4614f8bc9f59cbebb8b7f206" exitCode=0 Dec 09 10:30:21 crc kubenswrapper[4824]: I1209 10:30:21.603397 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh" event={"ID":"5c209e08-d978-483b-88b8-2f0ef683d62e","Type":"ContainerDied","Data":"733c1cfe764f6b4c0f5881cb31fd6c566319babe4614f8bc9f59cbebb8b7f206"} Dec 09 10:30:21 crc kubenswrapper[4824]: I1209 10:30:21.916960 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:30:21 crc kubenswrapper[4824]: E1209 10:30:21.922723 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.105982 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.266601 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rd6gv\" (UniqueName: \"kubernetes.io/projected/5c209e08-d978-483b-88b8-2f0ef683d62e-kube-api-access-rd6gv\") pod \"5c209e08-d978-483b-88b8-2f0ef683d62e\" (UID: \"5c209e08-d978-483b-88b8-2f0ef683d62e\") " Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.267153 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c209e08-d978-483b-88b8-2f0ef683d62e-inventory\") pod \"5c209e08-d978-483b-88b8-2f0ef683d62e\" (UID: \"5c209e08-d978-483b-88b8-2f0ef683d62e\") " Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.267231 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c209e08-d978-483b-88b8-2f0ef683d62e-ssh-key\") pod \"5c209e08-d978-483b-88b8-2f0ef683d62e\" (UID: \"5c209e08-d978-483b-88b8-2f0ef683d62e\") " Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.267369 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5c209e08-d978-483b-88b8-2f0ef683d62e-ovncontroller-config-0\") pod \"5c209e08-d978-483b-88b8-2f0ef683d62e\" (UID: \"5c209e08-d978-483b-88b8-2f0ef683d62e\") " Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.267566 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c209e08-d978-483b-88b8-2f0ef683d62e-ovn-combined-ca-bundle\") pod \"5c209e08-d978-483b-88b8-2f0ef683d62e\" (UID: \"5c209e08-d978-483b-88b8-2f0ef683d62e\") " Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.273252 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c209e08-d978-483b-88b8-2f0ef683d62e-kube-api-access-rd6gv" (OuterVolumeSpecName: "kube-api-access-rd6gv") pod "5c209e08-d978-483b-88b8-2f0ef683d62e" (UID: "5c209e08-d978-483b-88b8-2f0ef683d62e"). InnerVolumeSpecName "kube-api-access-rd6gv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.273772 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c209e08-d978-483b-88b8-2f0ef683d62e-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "5c209e08-d978-483b-88b8-2f0ef683d62e" (UID: "5c209e08-d978-483b-88b8-2f0ef683d62e"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.300407 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c209e08-d978-483b-88b8-2f0ef683d62e-inventory" (OuterVolumeSpecName: "inventory") pod "5c209e08-d978-483b-88b8-2f0ef683d62e" (UID: "5c209e08-d978-483b-88b8-2f0ef683d62e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.304091 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c209e08-d978-483b-88b8-2f0ef683d62e-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "5c209e08-d978-483b-88b8-2f0ef683d62e" (UID: "5c209e08-d978-483b-88b8-2f0ef683d62e"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.313684 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c209e08-d978-483b-88b8-2f0ef683d62e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5c209e08-d978-483b-88b8-2f0ef683d62e" (UID: "5c209e08-d978-483b-88b8-2f0ef683d62e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.376708 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c209e08-d978-483b-88b8-2f0ef683d62e-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.377037 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c209e08-d978-483b-88b8-2f0ef683d62e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.377178 4824 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5c209e08-d978-483b-88b8-2f0ef683d62e-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.377260 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c209e08-d978-483b-88b8-2f0ef683d62e-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.377402 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rd6gv\" (UniqueName: \"kubernetes.io/projected/5c209e08-d978-483b-88b8-2f0ef683d62e-kube-api-access-rd6gv\") on node \"crc\" DevicePath \"\"" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.634765 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh" event={"ID":"5c209e08-d978-483b-88b8-2f0ef683d62e","Type":"ContainerDied","Data":"01a734bc441edc2417175205a13060853085279d9f3353971b71afd2c6d3a735"} Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.635095 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01a734bc441edc2417175205a13060853085279d9f3353971b71afd2c6d3a735" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.635248 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dvksh" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.762583 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x"] Dec 09 10:30:23 crc kubenswrapper[4824]: E1209 10:30:23.764198 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="def49a3f-d880-44de-a0dc-3b57abd944a7" containerName="extract-content" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.765316 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="def49a3f-d880-44de-a0dc-3b57abd944a7" containerName="extract-content" Dec 09 10:30:23 crc kubenswrapper[4824]: E1209 10:30:23.765445 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c209e08-d978-483b-88b8-2f0ef683d62e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.765539 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c209e08-d978-483b-88b8-2f0ef683d62e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 09 10:30:23 crc kubenswrapper[4824]: E1209 10:30:23.765671 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="def49a3f-d880-44de-a0dc-3b57abd944a7" containerName="extract-utilities" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.765759 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="def49a3f-d880-44de-a0dc-3b57abd944a7" containerName="extract-utilities" Dec 09 10:30:23 crc kubenswrapper[4824]: E1209 10:30:23.765881 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="def49a3f-d880-44de-a0dc-3b57abd944a7" containerName="registry-server" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.765992 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="def49a3f-d880-44de-a0dc-3b57abd944a7" containerName="registry-server" Dec 09 10:30:23 crc kubenswrapper[4824]: E1209 10:30:23.766110 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65d6395b-1cca-479b-986c-49b2142e7a30" containerName="collect-profiles" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.766201 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="65d6395b-1cca-479b-986c-49b2142e7a30" containerName="collect-profiles" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.766855 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="def49a3f-d880-44de-a0dc-3b57abd944a7" containerName="registry-server" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.766982 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c209e08-d978-483b-88b8-2f0ef683d62e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.767101 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="65d6395b-1cca-479b-986c-49b2142e7a30" containerName="collect-profiles" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.768615 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.771184 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.771184 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.771401 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.771437 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.772720 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5nzld" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.780590 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.787076 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x"] Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.892387 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x\" (UID: \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.892774 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x\" (UID: \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.893049 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x\" (UID: \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.893253 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlqrc\" (UniqueName: \"kubernetes.io/projected/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-kube-api-access-qlqrc\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x\" (UID: \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.893328 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x\" (UID: \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.893416 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x\" (UID: \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.995287 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x\" (UID: \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.995352 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlqrc\" (UniqueName: \"kubernetes.io/projected/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-kube-api-access-qlqrc\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x\" (UID: \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.995385 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x\" (UID: \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.995431 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x\" (UID: \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.995482 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x\" (UID: \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" Dec 09 10:30:23 crc kubenswrapper[4824]: I1209 10:30:23.995532 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x\" (UID: \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" Dec 09 10:30:24 crc kubenswrapper[4824]: I1209 10:30:24.001169 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x\" (UID: \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" Dec 09 10:30:24 crc kubenswrapper[4824]: I1209 10:30:24.001368 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x\" (UID: \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" Dec 09 10:30:24 crc kubenswrapper[4824]: I1209 10:30:24.002891 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x\" (UID: \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" Dec 09 10:30:24 crc kubenswrapper[4824]: I1209 10:30:24.003607 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x\" (UID: \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" Dec 09 10:30:24 crc kubenswrapper[4824]: I1209 10:30:24.012049 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x\" (UID: \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" Dec 09 10:30:24 crc kubenswrapper[4824]: I1209 10:30:24.012989 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlqrc\" (UniqueName: \"kubernetes.io/projected/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-kube-api-access-qlqrc\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x\" (UID: \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" Dec 09 10:30:24 crc kubenswrapper[4824]: I1209 10:30:24.107368 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5nzld" Dec 09 10:30:24 crc kubenswrapper[4824]: I1209 10:30:24.114727 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" Dec 09 10:30:24 crc kubenswrapper[4824]: I1209 10:30:24.780301 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x"] Dec 09 10:30:25 crc kubenswrapper[4824]: I1209 10:30:25.211449 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 10:30:25 crc kubenswrapper[4824]: I1209 10:30:25.659163 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" event={"ID":"3fd9b8ef-292c-46ee-938e-62daa3bda9e0","Type":"ContainerStarted","Data":"f87cba1e64c3ba9fb6d994d655ff01f93aab3f088f09e1f22b695124af3222cf"} Dec 09 10:30:25 crc kubenswrapper[4824]: I1209 10:30:25.659512 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" event={"ID":"3fd9b8ef-292c-46ee-938e-62daa3bda9e0","Type":"ContainerStarted","Data":"44cae3566c3c9d27e43001c17023df3d3c3cd58511cb3da7b8cf43b7a59e0110"} Dec 09 10:30:25 crc kubenswrapper[4824]: I1209 10:30:25.687164 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" podStartSLOduration=2.270430458 podStartE2EDuration="2.687143121s" podCreationTimestamp="2025-12-09 10:30:23 +0000 UTC" firstStartedPulling="2025-12-09 10:30:24.792431261 +0000 UTC m=+2581.126935928" lastFinishedPulling="2025-12-09 10:30:25.209143924 +0000 UTC m=+2581.543648591" observedRunningTime="2025-12-09 10:30:25.682194976 +0000 UTC m=+2582.016699673" watchObservedRunningTime="2025-12-09 10:30:25.687143121 +0000 UTC m=+2582.021647788" Dec 09 10:30:32 crc kubenswrapper[4824]: I1209 10:30:32.910488 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:30:32 crc kubenswrapper[4824]: E1209 10:30:32.911321 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:30:44 crc kubenswrapper[4824]: I1209 10:30:44.921640 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:30:44 crc kubenswrapper[4824]: E1209 10:30:44.928051 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:30:56 crc kubenswrapper[4824]: I1209 10:30:56.911352 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:30:56 crc kubenswrapper[4824]: E1209 10:30:56.912245 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:31:07 crc kubenswrapper[4824]: I1209 10:31:07.912277 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:31:07 crc kubenswrapper[4824]: E1209 10:31:07.914490 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:31:16 crc kubenswrapper[4824]: I1209 10:31:16.275264 4824 generic.go:334] "Generic (PLEG): container finished" podID="3fd9b8ef-292c-46ee-938e-62daa3bda9e0" containerID="f87cba1e64c3ba9fb6d994d655ff01f93aab3f088f09e1f22b695124af3222cf" exitCode=0 Dec 09 10:31:16 crc kubenswrapper[4824]: I1209 10:31:16.275278 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" event={"ID":"3fd9b8ef-292c-46ee-938e-62daa3bda9e0","Type":"ContainerDied","Data":"f87cba1e64c3ba9fb6d994d655ff01f93aab3f088f09e1f22b695124af3222cf"} Dec 09 10:31:17 crc kubenswrapper[4824]: I1209 10:31:17.740405 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" Dec 09 10:31:17 crc kubenswrapper[4824]: I1209 10:31:17.886304 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-ssh-key\") pod \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\" (UID: \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\") " Dec 09 10:31:17 crc kubenswrapper[4824]: I1209 10:31:17.886399 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-neutron-ovn-metadata-agent-neutron-config-0\") pod \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\" (UID: \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\") " Dec 09 10:31:17 crc kubenswrapper[4824]: I1209 10:31:17.886539 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-nova-metadata-neutron-config-0\") pod \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\" (UID: \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\") " Dec 09 10:31:17 crc kubenswrapper[4824]: I1209 10:31:17.886652 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlqrc\" (UniqueName: \"kubernetes.io/projected/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-kube-api-access-qlqrc\") pod \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\" (UID: \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\") " Dec 09 10:31:17 crc kubenswrapper[4824]: I1209 10:31:17.886727 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-neutron-metadata-combined-ca-bundle\") pod \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\" (UID: \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\") " Dec 09 10:31:17 crc kubenswrapper[4824]: I1209 10:31:17.886838 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-inventory\") pod \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\" (UID: \"3fd9b8ef-292c-46ee-938e-62daa3bda9e0\") " Dec 09 10:31:17 crc kubenswrapper[4824]: I1209 10:31:17.891995 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-kube-api-access-qlqrc" (OuterVolumeSpecName: "kube-api-access-qlqrc") pod "3fd9b8ef-292c-46ee-938e-62daa3bda9e0" (UID: "3fd9b8ef-292c-46ee-938e-62daa3bda9e0"). InnerVolumeSpecName "kube-api-access-qlqrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:31:17 crc kubenswrapper[4824]: I1209 10:31:17.892282 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "3fd9b8ef-292c-46ee-938e-62daa3bda9e0" (UID: "3fd9b8ef-292c-46ee-938e-62daa3bda9e0"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:31:17 crc kubenswrapper[4824]: I1209 10:31:17.945441 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "3fd9b8ef-292c-46ee-938e-62daa3bda9e0" (UID: "3fd9b8ef-292c-46ee-938e-62daa3bda9e0"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:31:17 crc kubenswrapper[4824]: I1209 10:31:17.948051 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3fd9b8ef-292c-46ee-938e-62daa3bda9e0" (UID: "3fd9b8ef-292c-46ee-938e-62daa3bda9e0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:31:17 crc kubenswrapper[4824]: I1209 10:31:17.953550 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "3fd9b8ef-292c-46ee-938e-62daa3bda9e0" (UID: "3fd9b8ef-292c-46ee-938e-62daa3bda9e0"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:31:17 crc kubenswrapper[4824]: I1209 10:31:17.971712 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-inventory" (OuterVolumeSpecName: "inventory") pod "3fd9b8ef-292c-46ee-938e-62daa3bda9e0" (UID: "3fd9b8ef-292c-46ee-938e-62daa3bda9e0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:31:17 crc kubenswrapper[4824]: I1209 10:31:17.991044 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 10:31:17 crc kubenswrapper[4824]: I1209 10:31:17.991114 4824 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:31:17 crc kubenswrapper[4824]: I1209 10:31:17.991135 4824 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:31:17 crc kubenswrapper[4824]: I1209 10:31:17.991151 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlqrc\" (UniqueName: \"kubernetes.io/projected/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-kube-api-access-qlqrc\") on node \"crc\" DevicePath \"\"" Dec 09 10:31:17 crc kubenswrapper[4824]: I1209 10:31:17.991167 4824 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:31:17 crc kubenswrapper[4824]: I1209 10:31:17.991180 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3fd9b8ef-292c-46ee-938e-62daa3bda9e0-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.302352 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" event={"ID":"3fd9b8ef-292c-46ee-938e-62daa3bda9e0","Type":"ContainerDied","Data":"44cae3566c3c9d27e43001c17023df3d3c3cd58511cb3da7b8cf43b7a59e0110"} Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.302420 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44cae3566c3c9d27e43001c17023df3d3c3cd58511cb3da7b8cf43b7a59e0110" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.302493 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.417968 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk"] Dec 09 10:31:18 crc kubenswrapper[4824]: E1209 10:31:18.418909 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fd9b8ef-292c-46ee-938e-62daa3bda9e0" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.418929 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fd9b8ef-292c-46ee-938e-62daa3bda9e0" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.419282 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fd9b8ef-292c-46ee-938e-62daa3bda9e0" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.420503 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.423543 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5nzld" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.423942 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.423965 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.427123 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.427226 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.459454 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk"] Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.502331 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2df5cd6-d375-4545-bf5a-bd3df55fb736-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk\" (UID: \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.502567 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f2df5cd6-d375-4545-bf5a-bd3df55fb736-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk\" (UID: \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.502714 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8stlj\" (UniqueName: \"kubernetes.io/projected/f2df5cd6-d375-4545-bf5a-bd3df55fb736-kube-api-access-8stlj\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk\" (UID: \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.502765 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f2df5cd6-d375-4545-bf5a-bd3df55fb736-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk\" (UID: \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.503377 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f2df5cd6-d375-4545-bf5a-bd3df55fb736-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk\" (UID: \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.605359 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f2df5cd6-d375-4545-bf5a-bd3df55fb736-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk\" (UID: \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.605434 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2df5cd6-d375-4545-bf5a-bd3df55fb736-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk\" (UID: \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.605520 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f2df5cd6-d375-4545-bf5a-bd3df55fb736-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk\" (UID: \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.605562 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8stlj\" (UniqueName: \"kubernetes.io/projected/f2df5cd6-d375-4545-bf5a-bd3df55fb736-kube-api-access-8stlj\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk\" (UID: \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.605582 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f2df5cd6-d375-4545-bf5a-bd3df55fb736-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk\" (UID: \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.610666 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f2df5cd6-d375-4545-bf5a-bd3df55fb736-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk\" (UID: \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.611103 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f2df5cd6-d375-4545-bf5a-bd3df55fb736-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk\" (UID: \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.611684 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f2df5cd6-d375-4545-bf5a-bd3df55fb736-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk\" (UID: \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.612925 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2df5cd6-d375-4545-bf5a-bd3df55fb736-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk\" (UID: \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.626772 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8stlj\" (UniqueName: \"kubernetes.io/projected/f2df5cd6-d375-4545-bf5a-bd3df55fb736-kube-api-access-8stlj\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk\" (UID: \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.743539 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk" Dec 09 10:31:18 crc kubenswrapper[4824]: I1209 10:31:18.910898 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:31:18 crc kubenswrapper[4824]: E1209 10:31:18.911546 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:31:19 crc kubenswrapper[4824]: I1209 10:31:19.605056 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk"] Dec 09 10:31:20 crc kubenswrapper[4824]: I1209 10:31:20.326758 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk" event={"ID":"f2df5cd6-d375-4545-bf5a-bd3df55fb736","Type":"ContainerStarted","Data":"bb23925115311abbb2ad58046cd2574fc51fb0bfcc69cca6c33fa971b900fc73"} Dec 09 10:31:21 crc kubenswrapper[4824]: I1209 10:31:21.342582 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk" event={"ID":"f2df5cd6-d375-4545-bf5a-bd3df55fb736","Type":"ContainerStarted","Data":"b27c757bf38537806c90462214469fef3599dc14b29b1a0de5dc00642c35f1b2"} Dec 09 10:31:21 crc kubenswrapper[4824]: I1209 10:31:21.368105 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk" podStartSLOduration=2.86559192 podStartE2EDuration="3.36807234s" podCreationTimestamp="2025-12-09 10:31:18 +0000 UTC" firstStartedPulling="2025-12-09 10:31:19.606911343 +0000 UTC m=+2635.941416010" lastFinishedPulling="2025-12-09 10:31:20.109391763 +0000 UTC m=+2636.443896430" observedRunningTime="2025-12-09 10:31:21.366364066 +0000 UTC m=+2637.700868743" watchObservedRunningTime="2025-12-09 10:31:21.36807234 +0000 UTC m=+2637.702577007" Dec 09 10:31:30 crc kubenswrapper[4824]: I1209 10:31:30.911302 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:31:30 crc kubenswrapper[4824]: E1209 10:31:30.912156 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:31:42 crc kubenswrapper[4824]: I1209 10:31:42.911196 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:31:42 crc kubenswrapper[4824]: E1209 10:31:42.912052 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:31:55 crc kubenswrapper[4824]: I1209 10:31:55.911808 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:31:55 crc kubenswrapper[4824]: E1209 10:31:55.912486 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:32:06 crc kubenswrapper[4824]: I1209 10:32:06.911155 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:32:06 crc kubenswrapper[4824]: E1209 10:32:06.912097 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:32:12 crc kubenswrapper[4824]: I1209 10:32:12.725234 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4mttk"] Dec 09 10:32:12 crc kubenswrapper[4824]: I1209 10:32:12.731303 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4mttk" Dec 09 10:32:12 crc kubenswrapper[4824]: I1209 10:32:12.746282 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4mttk"] Dec 09 10:32:12 crc kubenswrapper[4824]: I1209 10:32:12.849500 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4088c9a4-bc39-40e8-a8d8-f35092178361-utilities\") pod \"redhat-operators-4mttk\" (UID: \"4088c9a4-bc39-40e8-a8d8-f35092178361\") " pod="openshift-marketplace/redhat-operators-4mttk" Dec 09 10:32:12 crc kubenswrapper[4824]: I1209 10:32:12.849741 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4088c9a4-bc39-40e8-a8d8-f35092178361-catalog-content\") pod \"redhat-operators-4mttk\" (UID: \"4088c9a4-bc39-40e8-a8d8-f35092178361\") " pod="openshift-marketplace/redhat-operators-4mttk" Dec 09 10:32:12 crc kubenswrapper[4824]: I1209 10:32:12.849770 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl5gn\" (UniqueName: \"kubernetes.io/projected/4088c9a4-bc39-40e8-a8d8-f35092178361-kube-api-access-wl5gn\") pod \"redhat-operators-4mttk\" (UID: \"4088c9a4-bc39-40e8-a8d8-f35092178361\") " pod="openshift-marketplace/redhat-operators-4mttk" Dec 09 10:32:12 crc kubenswrapper[4824]: I1209 10:32:12.951620 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4088c9a4-bc39-40e8-a8d8-f35092178361-utilities\") pod \"redhat-operators-4mttk\" (UID: \"4088c9a4-bc39-40e8-a8d8-f35092178361\") " pod="openshift-marketplace/redhat-operators-4mttk" Dec 09 10:32:12 crc kubenswrapper[4824]: I1209 10:32:12.951962 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4088c9a4-bc39-40e8-a8d8-f35092178361-catalog-content\") pod \"redhat-operators-4mttk\" (UID: \"4088c9a4-bc39-40e8-a8d8-f35092178361\") " pod="openshift-marketplace/redhat-operators-4mttk" Dec 09 10:32:12 crc kubenswrapper[4824]: I1209 10:32:12.952006 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl5gn\" (UniqueName: \"kubernetes.io/projected/4088c9a4-bc39-40e8-a8d8-f35092178361-kube-api-access-wl5gn\") pod \"redhat-operators-4mttk\" (UID: \"4088c9a4-bc39-40e8-a8d8-f35092178361\") " pod="openshift-marketplace/redhat-operators-4mttk" Dec 09 10:32:12 crc kubenswrapper[4824]: I1209 10:32:12.952287 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4088c9a4-bc39-40e8-a8d8-f35092178361-utilities\") pod \"redhat-operators-4mttk\" (UID: \"4088c9a4-bc39-40e8-a8d8-f35092178361\") " pod="openshift-marketplace/redhat-operators-4mttk" Dec 09 10:32:12 crc kubenswrapper[4824]: I1209 10:32:12.952902 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4088c9a4-bc39-40e8-a8d8-f35092178361-catalog-content\") pod \"redhat-operators-4mttk\" (UID: \"4088c9a4-bc39-40e8-a8d8-f35092178361\") " pod="openshift-marketplace/redhat-operators-4mttk" Dec 09 10:32:12 crc kubenswrapper[4824]: I1209 10:32:12.971716 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl5gn\" (UniqueName: \"kubernetes.io/projected/4088c9a4-bc39-40e8-a8d8-f35092178361-kube-api-access-wl5gn\") pod \"redhat-operators-4mttk\" (UID: \"4088c9a4-bc39-40e8-a8d8-f35092178361\") " pod="openshift-marketplace/redhat-operators-4mttk" Dec 09 10:32:13 crc kubenswrapper[4824]: I1209 10:32:13.062500 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4mttk" Dec 09 10:32:13 crc kubenswrapper[4824]: I1209 10:32:13.609038 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4mttk"] Dec 09 10:32:14 crc kubenswrapper[4824]: I1209 10:32:14.564116 4824 generic.go:334] "Generic (PLEG): container finished" podID="4088c9a4-bc39-40e8-a8d8-f35092178361" containerID="ced33d168c5e30b8ec59568aa5db78e40ed78ed13822b10b09db21991b8a1445" exitCode=0 Dec 09 10:32:14 crc kubenswrapper[4824]: I1209 10:32:14.564192 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4mttk" event={"ID":"4088c9a4-bc39-40e8-a8d8-f35092178361","Type":"ContainerDied","Data":"ced33d168c5e30b8ec59568aa5db78e40ed78ed13822b10b09db21991b8a1445"} Dec 09 10:32:14 crc kubenswrapper[4824]: I1209 10:32:14.564395 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4mttk" event={"ID":"4088c9a4-bc39-40e8-a8d8-f35092178361","Type":"ContainerStarted","Data":"a2c48f65d8fd793128c2f3a587b1794b13b91359390bcc5187ed8d0e11ba57dd"} Dec 09 10:32:15 crc kubenswrapper[4824]: I1209 10:32:15.577448 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4mttk" event={"ID":"4088c9a4-bc39-40e8-a8d8-f35092178361","Type":"ContainerStarted","Data":"1afeb0b92149809922645908e5a370d75ce9f919bad157b47e5f922c20e08e23"} Dec 09 10:32:19 crc kubenswrapper[4824]: I1209 10:32:19.638566 4824 generic.go:334] "Generic (PLEG): container finished" podID="4088c9a4-bc39-40e8-a8d8-f35092178361" containerID="1afeb0b92149809922645908e5a370d75ce9f919bad157b47e5f922c20e08e23" exitCode=0 Dec 09 10:32:19 crc kubenswrapper[4824]: I1209 10:32:19.638986 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4mttk" event={"ID":"4088c9a4-bc39-40e8-a8d8-f35092178361","Type":"ContainerDied","Data":"1afeb0b92149809922645908e5a370d75ce9f919bad157b47e5f922c20e08e23"} Dec 09 10:32:20 crc kubenswrapper[4824]: I1209 10:32:20.651643 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4mttk" event={"ID":"4088c9a4-bc39-40e8-a8d8-f35092178361","Type":"ContainerStarted","Data":"cebb3f532b76e7367f64bf75becdfe5e675aa30de6eda7c69ea6694848e53d24"} Dec 09 10:32:20 crc kubenswrapper[4824]: I1209 10:32:20.674040 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4mttk" podStartSLOduration=3.216505374 podStartE2EDuration="8.67402058s" podCreationTimestamp="2025-12-09 10:32:12 +0000 UTC" firstStartedPulling="2025-12-09 10:32:14.568290961 +0000 UTC m=+2690.902795628" lastFinishedPulling="2025-12-09 10:32:20.025806167 +0000 UTC m=+2696.360310834" observedRunningTime="2025-12-09 10:32:20.67018943 +0000 UTC m=+2697.004694097" watchObservedRunningTime="2025-12-09 10:32:20.67402058 +0000 UTC m=+2697.008525247" Dec 09 10:32:20 crc kubenswrapper[4824]: I1209 10:32:20.910635 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:32:20 crc kubenswrapper[4824]: E1209 10:32:20.911020 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:32:23 crc kubenswrapper[4824]: I1209 10:32:23.064241 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4mttk" Dec 09 10:32:23 crc kubenswrapper[4824]: I1209 10:32:23.064887 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4mttk" Dec 09 10:32:24 crc kubenswrapper[4824]: I1209 10:32:24.117993 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4mttk" podUID="4088c9a4-bc39-40e8-a8d8-f35092178361" containerName="registry-server" probeResult="failure" output=< Dec 09 10:32:24 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 10:32:24 crc kubenswrapper[4824]: > Dec 09 10:32:33 crc kubenswrapper[4824]: I1209 10:32:33.115061 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4mttk" Dec 09 10:32:33 crc kubenswrapper[4824]: I1209 10:32:33.168218 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4mttk" Dec 09 10:32:33 crc kubenswrapper[4824]: I1209 10:32:33.354904 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4mttk"] Dec 09 10:32:34 crc kubenswrapper[4824]: I1209 10:32:34.806721 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4mttk" podUID="4088c9a4-bc39-40e8-a8d8-f35092178361" containerName="registry-server" containerID="cri-o://cebb3f532b76e7367f64bf75becdfe5e675aa30de6eda7c69ea6694848e53d24" gracePeriod=2 Dec 09 10:32:34 crc kubenswrapper[4824]: I1209 10:32:34.910654 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:32:35 crc kubenswrapper[4824]: I1209 10:32:35.405022 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4mttk" Dec 09 10:32:35 crc kubenswrapper[4824]: I1209 10:32:35.601372 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wl5gn\" (UniqueName: \"kubernetes.io/projected/4088c9a4-bc39-40e8-a8d8-f35092178361-kube-api-access-wl5gn\") pod \"4088c9a4-bc39-40e8-a8d8-f35092178361\" (UID: \"4088c9a4-bc39-40e8-a8d8-f35092178361\") " Dec 09 10:32:35 crc kubenswrapper[4824]: I1209 10:32:35.602034 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4088c9a4-bc39-40e8-a8d8-f35092178361-catalog-content\") pod \"4088c9a4-bc39-40e8-a8d8-f35092178361\" (UID: \"4088c9a4-bc39-40e8-a8d8-f35092178361\") " Dec 09 10:32:35 crc kubenswrapper[4824]: I1209 10:32:35.602269 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4088c9a4-bc39-40e8-a8d8-f35092178361-utilities\") pod \"4088c9a4-bc39-40e8-a8d8-f35092178361\" (UID: \"4088c9a4-bc39-40e8-a8d8-f35092178361\") " Dec 09 10:32:35 crc kubenswrapper[4824]: I1209 10:32:35.602964 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4088c9a4-bc39-40e8-a8d8-f35092178361-utilities" (OuterVolumeSpecName: "utilities") pod "4088c9a4-bc39-40e8-a8d8-f35092178361" (UID: "4088c9a4-bc39-40e8-a8d8-f35092178361"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:32:35 crc kubenswrapper[4824]: I1209 10:32:35.603391 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4088c9a4-bc39-40e8-a8d8-f35092178361-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 10:32:35 crc kubenswrapper[4824]: I1209 10:32:35.617629 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4088c9a4-bc39-40e8-a8d8-f35092178361-kube-api-access-wl5gn" (OuterVolumeSpecName: "kube-api-access-wl5gn") pod "4088c9a4-bc39-40e8-a8d8-f35092178361" (UID: "4088c9a4-bc39-40e8-a8d8-f35092178361"). InnerVolumeSpecName "kube-api-access-wl5gn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:32:35 crc kubenswrapper[4824]: I1209 10:32:35.706277 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wl5gn\" (UniqueName: \"kubernetes.io/projected/4088c9a4-bc39-40e8-a8d8-f35092178361-kube-api-access-wl5gn\") on node \"crc\" DevicePath \"\"" Dec 09 10:32:35 crc kubenswrapper[4824]: I1209 10:32:35.717219 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4088c9a4-bc39-40e8-a8d8-f35092178361-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4088c9a4-bc39-40e8-a8d8-f35092178361" (UID: "4088c9a4-bc39-40e8-a8d8-f35092178361"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:32:35 crc kubenswrapper[4824]: I1209 10:32:35.808209 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4088c9a4-bc39-40e8-a8d8-f35092178361-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 10:32:35 crc kubenswrapper[4824]: I1209 10:32:35.826587 4824 generic.go:334] "Generic (PLEG): container finished" podID="4088c9a4-bc39-40e8-a8d8-f35092178361" containerID="cebb3f532b76e7367f64bf75becdfe5e675aa30de6eda7c69ea6694848e53d24" exitCode=0 Dec 09 10:32:35 crc kubenswrapper[4824]: I1209 10:32:35.826698 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4mttk" event={"ID":"4088c9a4-bc39-40e8-a8d8-f35092178361","Type":"ContainerDied","Data":"cebb3f532b76e7367f64bf75becdfe5e675aa30de6eda7c69ea6694848e53d24"} Dec 09 10:32:35 crc kubenswrapper[4824]: I1209 10:32:35.826745 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4mttk" event={"ID":"4088c9a4-bc39-40e8-a8d8-f35092178361","Type":"ContainerDied","Data":"a2c48f65d8fd793128c2f3a587b1794b13b91359390bcc5187ed8d0e11ba57dd"} Dec 09 10:32:35 crc kubenswrapper[4824]: I1209 10:32:35.826818 4824 scope.go:117] "RemoveContainer" containerID="cebb3f532b76e7367f64bf75becdfe5e675aa30de6eda7c69ea6694848e53d24" Dec 09 10:32:35 crc kubenswrapper[4824]: I1209 10:32:35.827057 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4mttk" Dec 09 10:32:35 crc kubenswrapper[4824]: I1209 10:32:35.835861 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerStarted","Data":"b963f4bb52756cf9a0c69e35abd4f8d2f8fba82a3075a961f196fd16bdeaec5e"} Dec 09 10:32:35 crc kubenswrapper[4824]: I1209 10:32:35.912609 4824 scope.go:117] "RemoveContainer" containerID="1afeb0b92149809922645908e5a370d75ce9f919bad157b47e5f922c20e08e23" Dec 09 10:32:35 crc kubenswrapper[4824]: I1209 10:32:35.955680 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4mttk"] Dec 09 10:32:35 crc kubenswrapper[4824]: I1209 10:32:35.955719 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4mttk"] Dec 09 10:32:35 crc kubenswrapper[4824]: I1209 10:32:35.958142 4824 scope.go:117] "RemoveContainer" containerID="ced33d168c5e30b8ec59568aa5db78e40ed78ed13822b10b09db21991b8a1445" Dec 09 10:32:36 crc kubenswrapper[4824]: I1209 10:32:36.015341 4824 scope.go:117] "RemoveContainer" containerID="cebb3f532b76e7367f64bf75becdfe5e675aa30de6eda7c69ea6694848e53d24" Dec 09 10:32:36 crc kubenswrapper[4824]: E1209 10:32:36.016290 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cebb3f532b76e7367f64bf75becdfe5e675aa30de6eda7c69ea6694848e53d24\": container with ID starting with cebb3f532b76e7367f64bf75becdfe5e675aa30de6eda7c69ea6694848e53d24 not found: ID does not exist" containerID="cebb3f532b76e7367f64bf75becdfe5e675aa30de6eda7c69ea6694848e53d24" Dec 09 10:32:36 crc kubenswrapper[4824]: I1209 10:32:36.016335 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cebb3f532b76e7367f64bf75becdfe5e675aa30de6eda7c69ea6694848e53d24"} err="failed to get container status \"cebb3f532b76e7367f64bf75becdfe5e675aa30de6eda7c69ea6694848e53d24\": rpc error: code = NotFound desc = could not find container \"cebb3f532b76e7367f64bf75becdfe5e675aa30de6eda7c69ea6694848e53d24\": container with ID starting with cebb3f532b76e7367f64bf75becdfe5e675aa30de6eda7c69ea6694848e53d24 not found: ID does not exist" Dec 09 10:32:36 crc kubenswrapper[4824]: I1209 10:32:36.016369 4824 scope.go:117] "RemoveContainer" containerID="1afeb0b92149809922645908e5a370d75ce9f919bad157b47e5f922c20e08e23" Dec 09 10:32:36 crc kubenswrapper[4824]: E1209 10:32:36.016965 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1afeb0b92149809922645908e5a370d75ce9f919bad157b47e5f922c20e08e23\": container with ID starting with 1afeb0b92149809922645908e5a370d75ce9f919bad157b47e5f922c20e08e23 not found: ID does not exist" containerID="1afeb0b92149809922645908e5a370d75ce9f919bad157b47e5f922c20e08e23" Dec 09 10:32:36 crc kubenswrapper[4824]: I1209 10:32:36.017049 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1afeb0b92149809922645908e5a370d75ce9f919bad157b47e5f922c20e08e23"} err="failed to get container status \"1afeb0b92149809922645908e5a370d75ce9f919bad157b47e5f922c20e08e23\": rpc error: code = NotFound desc = could not find container \"1afeb0b92149809922645908e5a370d75ce9f919bad157b47e5f922c20e08e23\": container with ID starting with 1afeb0b92149809922645908e5a370d75ce9f919bad157b47e5f922c20e08e23 not found: ID does not exist" Dec 09 10:32:36 crc kubenswrapper[4824]: I1209 10:32:36.017159 4824 scope.go:117] "RemoveContainer" containerID="ced33d168c5e30b8ec59568aa5db78e40ed78ed13822b10b09db21991b8a1445" Dec 09 10:32:36 crc kubenswrapper[4824]: E1209 10:32:36.017957 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ced33d168c5e30b8ec59568aa5db78e40ed78ed13822b10b09db21991b8a1445\": container with ID starting with ced33d168c5e30b8ec59568aa5db78e40ed78ed13822b10b09db21991b8a1445 not found: ID does not exist" containerID="ced33d168c5e30b8ec59568aa5db78e40ed78ed13822b10b09db21991b8a1445" Dec 09 10:32:36 crc kubenswrapper[4824]: I1209 10:32:36.018006 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ced33d168c5e30b8ec59568aa5db78e40ed78ed13822b10b09db21991b8a1445"} err="failed to get container status \"ced33d168c5e30b8ec59568aa5db78e40ed78ed13822b10b09db21991b8a1445\": rpc error: code = NotFound desc = could not find container \"ced33d168c5e30b8ec59568aa5db78e40ed78ed13822b10b09db21991b8a1445\": container with ID starting with ced33d168c5e30b8ec59568aa5db78e40ed78ed13822b10b09db21991b8a1445 not found: ID does not exist" Dec 09 10:32:37 crc kubenswrapper[4824]: I1209 10:32:37.923452 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4088c9a4-bc39-40e8-a8d8-f35092178361" path="/var/lib/kubelet/pods/4088c9a4-bc39-40e8-a8d8-f35092178361/volumes" Dec 09 10:34:31 crc kubenswrapper[4824]: I1209 10:34:31.885908 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-w5t64"] Dec 09 10:34:31 crc kubenswrapper[4824]: E1209 10:34:31.887059 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4088c9a4-bc39-40e8-a8d8-f35092178361" containerName="extract-content" Dec 09 10:34:31 crc kubenswrapper[4824]: I1209 10:34:31.887074 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4088c9a4-bc39-40e8-a8d8-f35092178361" containerName="extract-content" Dec 09 10:34:31 crc kubenswrapper[4824]: E1209 10:34:31.887087 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4088c9a4-bc39-40e8-a8d8-f35092178361" containerName="extract-utilities" Dec 09 10:34:31 crc kubenswrapper[4824]: I1209 10:34:31.887094 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4088c9a4-bc39-40e8-a8d8-f35092178361" containerName="extract-utilities" Dec 09 10:34:31 crc kubenswrapper[4824]: E1209 10:34:31.887119 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4088c9a4-bc39-40e8-a8d8-f35092178361" containerName="registry-server" Dec 09 10:34:31 crc kubenswrapper[4824]: I1209 10:34:31.887127 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4088c9a4-bc39-40e8-a8d8-f35092178361" containerName="registry-server" Dec 09 10:34:31 crc kubenswrapper[4824]: I1209 10:34:31.887343 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4088c9a4-bc39-40e8-a8d8-f35092178361" containerName="registry-server" Dec 09 10:34:31 crc kubenswrapper[4824]: I1209 10:34:31.889543 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w5t64" Dec 09 10:34:31 crc kubenswrapper[4824]: I1209 10:34:31.899590 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w5t64"] Dec 09 10:34:32 crc kubenswrapper[4824]: I1209 10:34:32.013372 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl6nb\" (UniqueName: \"kubernetes.io/projected/7c5db1b2-abf3-4874-a838-ad652a7be779-kube-api-access-kl6nb\") pod \"community-operators-w5t64\" (UID: \"7c5db1b2-abf3-4874-a838-ad652a7be779\") " pod="openshift-marketplace/community-operators-w5t64" Dec 09 10:34:32 crc kubenswrapper[4824]: I1209 10:34:32.014710 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c5db1b2-abf3-4874-a838-ad652a7be779-catalog-content\") pod \"community-operators-w5t64\" (UID: \"7c5db1b2-abf3-4874-a838-ad652a7be779\") " pod="openshift-marketplace/community-operators-w5t64" Dec 09 10:34:32 crc kubenswrapper[4824]: I1209 10:34:32.014926 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c5db1b2-abf3-4874-a838-ad652a7be779-utilities\") pod \"community-operators-w5t64\" (UID: \"7c5db1b2-abf3-4874-a838-ad652a7be779\") " pod="openshift-marketplace/community-operators-w5t64" Dec 09 10:34:32 crc kubenswrapper[4824]: I1209 10:34:32.117552 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c5db1b2-abf3-4874-a838-ad652a7be779-utilities\") pod \"community-operators-w5t64\" (UID: \"7c5db1b2-abf3-4874-a838-ad652a7be779\") " pod="openshift-marketplace/community-operators-w5t64" Dec 09 10:34:32 crc kubenswrapper[4824]: I1209 10:34:32.118007 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl6nb\" (UniqueName: \"kubernetes.io/projected/7c5db1b2-abf3-4874-a838-ad652a7be779-kube-api-access-kl6nb\") pod \"community-operators-w5t64\" (UID: \"7c5db1b2-abf3-4874-a838-ad652a7be779\") " pod="openshift-marketplace/community-operators-w5t64" Dec 09 10:34:32 crc kubenswrapper[4824]: I1209 10:34:32.118248 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c5db1b2-abf3-4874-a838-ad652a7be779-catalog-content\") pod \"community-operators-w5t64\" (UID: \"7c5db1b2-abf3-4874-a838-ad652a7be779\") " pod="openshift-marketplace/community-operators-w5t64" Dec 09 10:34:32 crc kubenswrapper[4824]: I1209 10:34:32.118958 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c5db1b2-abf3-4874-a838-ad652a7be779-catalog-content\") pod \"community-operators-w5t64\" (UID: \"7c5db1b2-abf3-4874-a838-ad652a7be779\") " pod="openshift-marketplace/community-operators-w5t64" Dec 09 10:34:32 crc kubenswrapper[4824]: I1209 10:34:32.119271 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c5db1b2-abf3-4874-a838-ad652a7be779-utilities\") pod \"community-operators-w5t64\" (UID: \"7c5db1b2-abf3-4874-a838-ad652a7be779\") " pod="openshift-marketplace/community-operators-w5t64" Dec 09 10:34:32 crc kubenswrapper[4824]: I1209 10:34:32.146312 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl6nb\" (UniqueName: \"kubernetes.io/projected/7c5db1b2-abf3-4874-a838-ad652a7be779-kube-api-access-kl6nb\") pod \"community-operators-w5t64\" (UID: \"7c5db1b2-abf3-4874-a838-ad652a7be779\") " pod="openshift-marketplace/community-operators-w5t64" Dec 09 10:34:32 crc kubenswrapper[4824]: I1209 10:34:32.223633 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w5t64" Dec 09 10:34:32 crc kubenswrapper[4824]: I1209 10:34:32.853895 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w5t64"] Dec 09 10:34:33 crc kubenswrapper[4824]: I1209 10:34:33.201634 4824 generic.go:334] "Generic (PLEG): container finished" podID="7c5db1b2-abf3-4874-a838-ad652a7be779" containerID="dbbf1eb68f5349814265da24ae192e2631f9af4b41900e7c2af383c78683e47d" exitCode=0 Dec 09 10:34:33 crc kubenswrapper[4824]: I1209 10:34:33.201699 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w5t64" event={"ID":"7c5db1b2-abf3-4874-a838-ad652a7be779","Type":"ContainerDied","Data":"dbbf1eb68f5349814265da24ae192e2631f9af4b41900e7c2af383c78683e47d"} Dec 09 10:34:33 crc kubenswrapper[4824]: I1209 10:34:33.201923 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w5t64" event={"ID":"7c5db1b2-abf3-4874-a838-ad652a7be779","Type":"ContainerStarted","Data":"31bb0a954962105d16d7e4ce2d8b8b2addf6ccc5900145e04f65819f4e6208b4"} Dec 09 10:34:33 crc kubenswrapper[4824]: I1209 10:34:33.204436 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 10:34:35 crc kubenswrapper[4824]: I1209 10:34:35.383138 4824 generic.go:334] "Generic (PLEG): container finished" podID="7c5db1b2-abf3-4874-a838-ad652a7be779" containerID="87d192e8516b1afa2293cb69dccd8c50ec2f16a8e6ce32545f33a59901ea193f" exitCode=0 Dec 09 10:34:35 crc kubenswrapper[4824]: I1209 10:34:35.385181 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w5t64" event={"ID":"7c5db1b2-abf3-4874-a838-ad652a7be779","Type":"ContainerDied","Data":"87d192e8516b1afa2293cb69dccd8c50ec2f16a8e6ce32545f33a59901ea193f"} Dec 09 10:34:36 crc kubenswrapper[4824]: I1209 10:34:36.396371 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w5t64" event={"ID":"7c5db1b2-abf3-4874-a838-ad652a7be779","Type":"ContainerStarted","Data":"dd16512548db40fa3a3c98f7428823957ea92ed172c98c10c2558c2a1b7d0a62"} Dec 09 10:34:36 crc kubenswrapper[4824]: I1209 10:34:36.431672 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-w5t64" podStartSLOduration=2.789627319 podStartE2EDuration="5.431619537s" podCreationTimestamp="2025-12-09 10:34:31 +0000 UTC" firstStartedPulling="2025-12-09 10:34:33.204048641 +0000 UTC m=+2829.538553328" lastFinishedPulling="2025-12-09 10:34:35.846040879 +0000 UTC m=+2832.180545546" observedRunningTime="2025-12-09 10:34:36.413196428 +0000 UTC m=+2832.747701115" watchObservedRunningTime="2025-12-09 10:34:36.431619537 +0000 UTC m=+2832.766124204" Dec 09 10:34:42 crc kubenswrapper[4824]: I1209 10:34:42.224950 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-w5t64" Dec 09 10:34:42 crc kubenswrapper[4824]: I1209 10:34:42.225568 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-w5t64" Dec 09 10:34:42 crc kubenswrapper[4824]: I1209 10:34:42.289285 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-w5t64" Dec 09 10:34:42 crc kubenswrapper[4824]: I1209 10:34:42.515280 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-w5t64" Dec 09 10:34:42 crc kubenswrapper[4824]: I1209 10:34:42.569706 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w5t64"] Dec 09 10:34:44 crc kubenswrapper[4824]: I1209 10:34:44.725592 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-w5t64" podUID="7c5db1b2-abf3-4874-a838-ad652a7be779" containerName="registry-server" containerID="cri-o://dd16512548db40fa3a3c98f7428823957ea92ed172c98c10c2558c2a1b7d0a62" gracePeriod=2 Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.293648 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w5t64" Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.432381 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c5db1b2-abf3-4874-a838-ad652a7be779-utilities\") pod \"7c5db1b2-abf3-4874-a838-ad652a7be779\" (UID: \"7c5db1b2-abf3-4874-a838-ad652a7be779\") " Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.432527 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kl6nb\" (UniqueName: \"kubernetes.io/projected/7c5db1b2-abf3-4874-a838-ad652a7be779-kube-api-access-kl6nb\") pod \"7c5db1b2-abf3-4874-a838-ad652a7be779\" (UID: \"7c5db1b2-abf3-4874-a838-ad652a7be779\") " Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.432613 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c5db1b2-abf3-4874-a838-ad652a7be779-catalog-content\") pod \"7c5db1b2-abf3-4874-a838-ad652a7be779\" (UID: \"7c5db1b2-abf3-4874-a838-ad652a7be779\") " Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.433686 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c5db1b2-abf3-4874-a838-ad652a7be779-utilities" (OuterVolumeSpecName: "utilities") pod "7c5db1b2-abf3-4874-a838-ad652a7be779" (UID: "7c5db1b2-abf3-4874-a838-ad652a7be779"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.450042 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c5db1b2-abf3-4874-a838-ad652a7be779-kube-api-access-kl6nb" (OuterVolumeSpecName: "kube-api-access-kl6nb") pod "7c5db1b2-abf3-4874-a838-ad652a7be779" (UID: "7c5db1b2-abf3-4874-a838-ad652a7be779"). InnerVolumeSpecName "kube-api-access-kl6nb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.514636 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c5db1b2-abf3-4874-a838-ad652a7be779-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c5db1b2-abf3-4874-a838-ad652a7be779" (UID: "7c5db1b2-abf3-4874-a838-ad652a7be779"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.542339 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c5db1b2-abf3-4874-a838-ad652a7be779-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.542385 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kl6nb\" (UniqueName: \"kubernetes.io/projected/7c5db1b2-abf3-4874-a838-ad652a7be779-kube-api-access-kl6nb\") on node \"crc\" DevicePath \"\"" Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.542401 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c5db1b2-abf3-4874-a838-ad652a7be779-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.738110 4824 generic.go:334] "Generic (PLEG): container finished" podID="7c5db1b2-abf3-4874-a838-ad652a7be779" containerID="dd16512548db40fa3a3c98f7428823957ea92ed172c98c10c2558c2a1b7d0a62" exitCode=0 Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.738162 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w5t64" event={"ID":"7c5db1b2-abf3-4874-a838-ad652a7be779","Type":"ContainerDied","Data":"dd16512548db40fa3a3c98f7428823957ea92ed172c98c10c2558c2a1b7d0a62"} Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.738192 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w5t64" Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.738204 4824 scope.go:117] "RemoveContainer" containerID="dd16512548db40fa3a3c98f7428823957ea92ed172c98c10c2558c2a1b7d0a62" Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.738193 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w5t64" event={"ID":"7c5db1b2-abf3-4874-a838-ad652a7be779","Type":"ContainerDied","Data":"31bb0a954962105d16d7e4ce2d8b8b2addf6ccc5900145e04f65819f4e6208b4"} Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.758742 4824 scope.go:117] "RemoveContainer" containerID="87d192e8516b1afa2293cb69dccd8c50ec2f16a8e6ce32545f33a59901ea193f" Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.774516 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w5t64"] Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.786254 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-w5t64"] Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.793828 4824 scope.go:117] "RemoveContainer" containerID="dbbf1eb68f5349814265da24ae192e2631f9af4b41900e7c2af383c78683e47d" Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.838544 4824 scope.go:117] "RemoveContainer" containerID="dd16512548db40fa3a3c98f7428823957ea92ed172c98c10c2558c2a1b7d0a62" Dec 09 10:34:45 crc kubenswrapper[4824]: E1209 10:34:45.839042 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd16512548db40fa3a3c98f7428823957ea92ed172c98c10c2558c2a1b7d0a62\": container with ID starting with dd16512548db40fa3a3c98f7428823957ea92ed172c98c10c2558c2a1b7d0a62 not found: ID does not exist" containerID="dd16512548db40fa3a3c98f7428823957ea92ed172c98c10c2558c2a1b7d0a62" Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.839111 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd16512548db40fa3a3c98f7428823957ea92ed172c98c10c2558c2a1b7d0a62"} err="failed to get container status \"dd16512548db40fa3a3c98f7428823957ea92ed172c98c10c2558c2a1b7d0a62\": rpc error: code = NotFound desc = could not find container \"dd16512548db40fa3a3c98f7428823957ea92ed172c98c10c2558c2a1b7d0a62\": container with ID starting with dd16512548db40fa3a3c98f7428823957ea92ed172c98c10c2558c2a1b7d0a62 not found: ID does not exist" Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.839140 4824 scope.go:117] "RemoveContainer" containerID="87d192e8516b1afa2293cb69dccd8c50ec2f16a8e6ce32545f33a59901ea193f" Dec 09 10:34:45 crc kubenswrapper[4824]: E1209 10:34:45.839435 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87d192e8516b1afa2293cb69dccd8c50ec2f16a8e6ce32545f33a59901ea193f\": container with ID starting with 87d192e8516b1afa2293cb69dccd8c50ec2f16a8e6ce32545f33a59901ea193f not found: ID does not exist" containerID="87d192e8516b1afa2293cb69dccd8c50ec2f16a8e6ce32545f33a59901ea193f" Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.839464 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87d192e8516b1afa2293cb69dccd8c50ec2f16a8e6ce32545f33a59901ea193f"} err="failed to get container status \"87d192e8516b1afa2293cb69dccd8c50ec2f16a8e6ce32545f33a59901ea193f\": rpc error: code = NotFound desc = could not find container \"87d192e8516b1afa2293cb69dccd8c50ec2f16a8e6ce32545f33a59901ea193f\": container with ID starting with 87d192e8516b1afa2293cb69dccd8c50ec2f16a8e6ce32545f33a59901ea193f not found: ID does not exist" Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.839485 4824 scope.go:117] "RemoveContainer" containerID="dbbf1eb68f5349814265da24ae192e2631f9af4b41900e7c2af383c78683e47d" Dec 09 10:34:45 crc kubenswrapper[4824]: E1209 10:34:45.839765 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbbf1eb68f5349814265da24ae192e2631f9af4b41900e7c2af383c78683e47d\": container with ID starting with dbbf1eb68f5349814265da24ae192e2631f9af4b41900e7c2af383c78683e47d not found: ID does not exist" containerID="dbbf1eb68f5349814265da24ae192e2631f9af4b41900e7c2af383c78683e47d" Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.839839 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbbf1eb68f5349814265da24ae192e2631f9af4b41900e7c2af383c78683e47d"} err="failed to get container status \"dbbf1eb68f5349814265da24ae192e2631f9af4b41900e7c2af383c78683e47d\": rpc error: code = NotFound desc = could not find container \"dbbf1eb68f5349814265da24ae192e2631f9af4b41900e7c2af383c78683e47d\": container with ID starting with dbbf1eb68f5349814265da24ae192e2631f9af4b41900e7c2af383c78683e47d not found: ID does not exist" Dec 09 10:34:45 crc kubenswrapper[4824]: I1209 10:34:45.927062 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c5db1b2-abf3-4874-a838-ad652a7be779" path="/var/lib/kubelet/pods/7c5db1b2-abf3-4874-a838-ad652a7be779/volumes" Dec 09 10:35:02 crc kubenswrapper[4824]: I1209 10:35:02.861088 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:35:02 crc kubenswrapper[4824]: I1209 10:35:02.861555 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:35:32 crc kubenswrapper[4824]: I1209 10:35:32.860877 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:35:32 crc kubenswrapper[4824]: I1209 10:35:32.861341 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:36:02 crc kubenswrapper[4824]: I1209 10:36:02.861065 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:36:02 crc kubenswrapper[4824]: I1209 10:36:02.861845 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:36:02 crc kubenswrapper[4824]: I1209 10:36:02.861923 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 10:36:02 crc kubenswrapper[4824]: I1209 10:36:02.863284 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b963f4bb52756cf9a0c69e35abd4f8d2f8fba82a3075a961f196fd16bdeaec5e"} pod="openshift-machine-config-operator/machine-config-daemon-dth8x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 10:36:02 crc kubenswrapper[4824]: I1209 10:36:02.863370 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" containerID="cri-o://b963f4bb52756cf9a0c69e35abd4f8d2f8fba82a3075a961f196fd16bdeaec5e" gracePeriod=600 Dec 09 10:36:03 crc kubenswrapper[4824]: I1209 10:36:03.948249 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerID="b963f4bb52756cf9a0c69e35abd4f8d2f8fba82a3075a961f196fd16bdeaec5e" exitCode=0 Dec 09 10:36:03 crc kubenswrapper[4824]: I1209 10:36:03.948343 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerDied","Data":"b963f4bb52756cf9a0c69e35abd4f8d2f8fba82a3075a961f196fd16bdeaec5e"} Dec 09 10:36:03 crc kubenswrapper[4824]: I1209 10:36:03.948827 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerStarted","Data":"ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb"} Dec 09 10:36:03 crc kubenswrapper[4824]: I1209 10:36:03.948853 4824 scope.go:117] "RemoveContainer" containerID="2369cd8e14ae181ecdc6bd17765d52141363cc1fa3835ab02605f37c585a0008" Dec 09 10:36:12 crc kubenswrapper[4824]: I1209 10:36:12.051326 4824 generic.go:334] "Generic (PLEG): container finished" podID="f2df5cd6-d375-4545-bf5a-bd3df55fb736" containerID="b27c757bf38537806c90462214469fef3599dc14b29b1a0de5dc00642c35f1b2" exitCode=0 Dec 09 10:36:12 crc kubenswrapper[4824]: I1209 10:36:12.051395 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk" event={"ID":"f2df5cd6-d375-4545-bf5a-bd3df55fb736","Type":"ContainerDied","Data":"b27c757bf38537806c90462214469fef3599dc14b29b1a0de5dc00642c35f1b2"} Dec 09 10:36:13 crc kubenswrapper[4824]: I1209 10:36:13.592916 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk" Dec 09 10:36:13 crc kubenswrapper[4824]: I1209 10:36:13.636521 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f2df5cd6-d375-4545-bf5a-bd3df55fb736-libvirt-secret-0\") pod \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\" (UID: \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\") " Dec 09 10:36:13 crc kubenswrapper[4824]: I1209 10:36:13.636603 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f2df5cd6-d375-4545-bf5a-bd3df55fb736-ssh-key\") pod \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\" (UID: \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\") " Dec 09 10:36:13 crc kubenswrapper[4824]: I1209 10:36:13.637046 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f2df5cd6-d375-4545-bf5a-bd3df55fb736-inventory\") pod \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\" (UID: \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\") " Dec 09 10:36:13 crc kubenswrapper[4824]: I1209 10:36:13.637943 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8stlj\" (UniqueName: \"kubernetes.io/projected/f2df5cd6-d375-4545-bf5a-bd3df55fb736-kube-api-access-8stlj\") pod \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\" (UID: \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\") " Dec 09 10:36:13 crc kubenswrapper[4824]: I1209 10:36:13.638148 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2df5cd6-d375-4545-bf5a-bd3df55fb736-libvirt-combined-ca-bundle\") pod \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\" (UID: \"f2df5cd6-d375-4545-bf5a-bd3df55fb736\") " Dec 09 10:36:13 crc kubenswrapper[4824]: I1209 10:36:13.643859 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2df5cd6-d375-4545-bf5a-bd3df55fb736-kube-api-access-8stlj" (OuterVolumeSpecName: "kube-api-access-8stlj") pod "f2df5cd6-d375-4545-bf5a-bd3df55fb736" (UID: "f2df5cd6-d375-4545-bf5a-bd3df55fb736"). InnerVolumeSpecName "kube-api-access-8stlj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:36:13 crc kubenswrapper[4824]: I1209 10:36:13.643887 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2df5cd6-d375-4545-bf5a-bd3df55fb736-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "f2df5cd6-d375-4545-bf5a-bd3df55fb736" (UID: "f2df5cd6-d375-4545-bf5a-bd3df55fb736"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:36:13 crc kubenswrapper[4824]: I1209 10:36:13.673278 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2df5cd6-d375-4545-bf5a-bd3df55fb736-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "f2df5cd6-d375-4545-bf5a-bd3df55fb736" (UID: "f2df5cd6-d375-4545-bf5a-bd3df55fb736"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:36:13 crc kubenswrapper[4824]: I1209 10:36:13.674562 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2df5cd6-d375-4545-bf5a-bd3df55fb736-inventory" (OuterVolumeSpecName: "inventory") pod "f2df5cd6-d375-4545-bf5a-bd3df55fb736" (UID: "f2df5cd6-d375-4545-bf5a-bd3df55fb736"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:36:13 crc kubenswrapper[4824]: I1209 10:36:13.675056 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2df5cd6-d375-4545-bf5a-bd3df55fb736-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f2df5cd6-d375-4545-bf5a-bd3df55fb736" (UID: "f2df5cd6-d375-4545-bf5a-bd3df55fb736"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:36:13 crc kubenswrapper[4824]: I1209 10:36:13.741198 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f2df5cd6-d375-4545-bf5a-bd3df55fb736-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 10:36:13 crc kubenswrapper[4824]: I1209 10:36:13.741238 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8stlj\" (UniqueName: \"kubernetes.io/projected/f2df5cd6-d375-4545-bf5a-bd3df55fb736-kube-api-access-8stlj\") on node \"crc\" DevicePath \"\"" Dec 09 10:36:13 crc kubenswrapper[4824]: I1209 10:36:13.741250 4824 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2df5cd6-d375-4545-bf5a-bd3df55fb736-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:36:13 crc kubenswrapper[4824]: I1209 10:36:13.741262 4824 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f2df5cd6-d375-4545-bf5a-bd3df55fb736-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:36:13 crc kubenswrapper[4824]: I1209 10:36:13.741271 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f2df5cd6-d375-4545-bf5a-bd3df55fb736-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.072873 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk" event={"ID":"f2df5cd6-d375-4545-bf5a-bd3df55fb736","Type":"ContainerDied","Data":"bb23925115311abbb2ad58046cd2574fc51fb0bfcc69cca6c33fa971b900fc73"} Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.072919 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb23925115311abbb2ad58046cd2574fc51fb0bfcc69cca6c33fa971b900fc73" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.072954 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.171902 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms"] Dec 09 10:36:14 crc kubenswrapper[4824]: E1209 10:36:14.172873 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c5db1b2-abf3-4874-a838-ad652a7be779" containerName="extract-utilities" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.172902 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c5db1b2-abf3-4874-a838-ad652a7be779" containerName="extract-utilities" Dec 09 10:36:14 crc kubenswrapper[4824]: E1209 10:36:14.172979 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c5db1b2-abf3-4874-a838-ad652a7be779" containerName="registry-server" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.172989 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c5db1b2-abf3-4874-a838-ad652a7be779" containerName="registry-server" Dec 09 10:36:14 crc kubenswrapper[4824]: E1209 10:36:14.173005 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2df5cd6-d375-4545-bf5a-bd3df55fb736" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.173015 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2df5cd6-d375-4545-bf5a-bd3df55fb736" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 09 10:36:14 crc kubenswrapper[4824]: E1209 10:36:14.173037 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c5db1b2-abf3-4874-a838-ad652a7be779" containerName="extract-content" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.173044 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c5db1b2-abf3-4874-a838-ad652a7be779" containerName="extract-content" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.173330 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2df5cd6-d375-4545-bf5a-bd3df55fb736" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.173368 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c5db1b2-abf3-4874-a838-ad652a7be779" containerName="registry-server" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.174508 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.180454 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.180584 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.180660 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.180852 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.181435 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5nzld" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.181443 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.182108 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.190979 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms"] Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.252847 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.253014 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.253057 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.253196 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.253249 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.253327 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.253391 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.253424 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmslc\" (UniqueName: \"kubernetes.io/projected/facb5571-0efc-4b89-8f5a-69dcff002fa5-kube-api-access-vmslc\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.253455 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.354295 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.354374 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.354415 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.354480 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmslc\" (UniqueName: \"kubernetes.io/projected/facb5571-0efc-4b89-8f5a-69dcff002fa5-kube-api-access-vmslc\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.354517 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.354535 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.354622 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.354710 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.354754 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.356536 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.361480 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.361910 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.362005 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.362251 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.363221 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.363578 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.378538 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.383374 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmslc\" (UniqueName: \"kubernetes.io/projected/facb5571-0efc-4b89-8f5a-69dcff002fa5-kube-api-access-vmslc\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7v7ms\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:14 crc kubenswrapper[4824]: I1209 10:36:14.507696 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:36:15 crc kubenswrapper[4824]: I1209 10:36:15.086461 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms"] Dec 09 10:36:15 crc kubenswrapper[4824]: W1209 10:36:15.087908 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfacb5571_0efc_4b89_8f5a_69dcff002fa5.slice/crio-84d4a7ac7c8f3df79c222ca19e98f17015eb241c4ce66e2fa19203bdc01b55cc WatchSource:0}: Error finding container 84d4a7ac7c8f3df79c222ca19e98f17015eb241c4ce66e2fa19203bdc01b55cc: Status 404 returned error can't find the container with id 84d4a7ac7c8f3df79c222ca19e98f17015eb241c4ce66e2fa19203bdc01b55cc Dec 09 10:36:16 crc kubenswrapper[4824]: I1209 10:36:16.115509 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" event={"ID":"facb5571-0efc-4b89-8f5a-69dcff002fa5","Type":"ContainerStarted","Data":"4dd05322dcdc3ab0b6ea8db5dc231b1633d584fd70a4e4ec60ae5d2c8a0b2d2f"} Dec 09 10:36:16 crc kubenswrapper[4824]: I1209 10:36:16.115854 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" event={"ID":"facb5571-0efc-4b89-8f5a-69dcff002fa5","Type":"ContainerStarted","Data":"84d4a7ac7c8f3df79c222ca19e98f17015eb241c4ce66e2fa19203bdc01b55cc"} Dec 09 10:36:16 crc kubenswrapper[4824]: I1209 10:36:16.153359 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" podStartSLOduration=1.712877941 podStartE2EDuration="2.1533366s" podCreationTimestamp="2025-12-09 10:36:14 +0000 UTC" firstStartedPulling="2025-12-09 10:36:15.091078042 +0000 UTC m=+2931.425582709" lastFinishedPulling="2025-12-09 10:36:15.531536701 +0000 UTC m=+2931.866041368" observedRunningTime="2025-12-09 10:36:16.135494677 +0000 UTC m=+2932.469999354" watchObservedRunningTime="2025-12-09 10:36:16.1533366 +0000 UTC m=+2932.487841267" Dec 09 10:38:32 crc kubenswrapper[4824]: I1209 10:38:32.860901 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:38:32 crc kubenswrapper[4824]: I1209 10:38:32.861663 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:39:02 crc kubenswrapper[4824]: I1209 10:39:02.860571 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:39:02 crc kubenswrapper[4824]: I1209 10:39:02.861212 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:39:17 crc kubenswrapper[4824]: I1209 10:39:17.145940 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" event={"ID":"facb5571-0efc-4b89-8f5a-69dcff002fa5","Type":"ContainerDied","Data":"4dd05322dcdc3ab0b6ea8db5dc231b1633d584fd70a4e4ec60ae5d2c8a0b2d2f"} Dec 09 10:39:17 crc kubenswrapper[4824]: I1209 10:39:17.145778 4824 generic.go:334] "Generic (PLEG): container finished" podID="facb5571-0efc-4b89-8f5a-69dcff002fa5" containerID="4dd05322dcdc3ab0b6ea8db5dc231b1633d584fd70a4e4ec60ae5d2c8a0b2d2f" exitCode=0 Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.740729 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.855512 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmslc\" (UniqueName: \"kubernetes.io/projected/facb5571-0efc-4b89-8f5a-69dcff002fa5-kube-api-access-vmslc\") pod \"facb5571-0efc-4b89-8f5a-69dcff002fa5\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.855900 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-combined-ca-bundle\") pod \"facb5571-0efc-4b89-8f5a-69dcff002fa5\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.855988 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-ssh-key\") pod \"facb5571-0efc-4b89-8f5a-69dcff002fa5\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.856034 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-cell1-compute-config-1\") pod \"facb5571-0efc-4b89-8f5a-69dcff002fa5\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.856152 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-cell1-compute-config-0\") pod \"facb5571-0efc-4b89-8f5a-69dcff002fa5\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.856183 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-inventory\") pod \"facb5571-0efc-4b89-8f5a-69dcff002fa5\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.856226 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-migration-ssh-key-1\") pod \"facb5571-0efc-4b89-8f5a-69dcff002fa5\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.856252 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-extra-config-0\") pod \"facb5571-0efc-4b89-8f5a-69dcff002fa5\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.856339 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-migration-ssh-key-0\") pod \"facb5571-0efc-4b89-8f5a-69dcff002fa5\" (UID: \"facb5571-0efc-4b89-8f5a-69dcff002fa5\") " Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.864045 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "facb5571-0efc-4b89-8f5a-69dcff002fa5" (UID: "facb5571-0efc-4b89-8f5a-69dcff002fa5"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.871652 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/facb5571-0efc-4b89-8f5a-69dcff002fa5-kube-api-access-vmslc" (OuterVolumeSpecName: "kube-api-access-vmslc") pod "facb5571-0efc-4b89-8f5a-69dcff002fa5" (UID: "facb5571-0efc-4b89-8f5a-69dcff002fa5"). InnerVolumeSpecName "kube-api-access-vmslc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.890687 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "facb5571-0efc-4b89-8f5a-69dcff002fa5" (UID: "facb5571-0efc-4b89-8f5a-69dcff002fa5"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.891147 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-inventory" (OuterVolumeSpecName: "inventory") pod "facb5571-0efc-4b89-8f5a-69dcff002fa5" (UID: "facb5571-0efc-4b89-8f5a-69dcff002fa5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.897771 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "facb5571-0efc-4b89-8f5a-69dcff002fa5" (UID: "facb5571-0efc-4b89-8f5a-69dcff002fa5"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.898948 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "facb5571-0efc-4b89-8f5a-69dcff002fa5" (UID: "facb5571-0efc-4b89-8f5a-69dcff002fa5"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.905062 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "facb5571-0efc-4b89-8f5a-69dcff002fa5" (UID: "facb5571-0efc-4b89-8f5a-69dcff002fa5"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.917033 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "facb5571-0efc-4b89-8f5a-69dcff002fa5" (UID: "facb5571-0efc-4b89-8f5a-69dcff002fa5"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.928653 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "facb5571-0efc-4b89-8f5a-69dcff002fa5" (UID: "facb5571-0efc-4b89-8f5a-69dcff002fa5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.959772 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmslc\" (UniqueName: \"kubernetes.io/projected/facb5571-0efc-4b89-8f5a-69dcff002fa5-kube-api-access-vmslc\") on node \"crc\" DevicePath \"\"" Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.959820 4824 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.959830 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.959842 4824 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.959851 4824 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.959867 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.959876 4824 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.959885 4824 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:39:18 crc kubenswrapper[4824]: I1209 10:39:18.959893 4824 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/facb5571-0efc-4b89-8f5a-69dcff002fa5-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.177423 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" event={"ID":"facb5571-0efc-4b89-8f5a-69dcff002fa5","Type":"ContainerDied","Data":"84d4a7ac7c8f3df79c222ca19e98f17015eb241c4ce66e2fa19203bdc01b55cc"} Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.177500 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84d4a7ac7c8f3df79c222ca19e98f17015eb241c4ce66e2fa19203bdc01b55cc" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.178075 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7v7ms" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.324497 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl"] Dec 09 10:39:19 crc kubenswrapper[4824]: E1209 10:39:19.325347 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="facb5571-0efc-4b89-8f5a-69dcff002fa5" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.325381 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="facb5571-0efc-4b89-8f5a-69dcff002fa5" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.325770 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="facb5571-0efc-4b89-8f5a-69dcff002fa5" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.326949 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.329329 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.334675 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5nzld" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.334720 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.334866 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.334912 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.339001 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl"] Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.472850 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-grqfl\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.472940 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-grqfl\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.472996 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-grqfl\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.473022 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-grqfl\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.473044 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-grqfl\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.473102 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-grqfl\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.473155 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7ttz\" (UniqueName: \"kubernetes.io/projected/86bfd02c-d226-48e8-bace-e961440bdb43-kube-api-access-x7ttz\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-grqfl\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.575670 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-grqfl\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.575751 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-grqfl\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.575798 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-grqfl\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.575830 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-grqfl\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.575888 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-grqfl\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.575935 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7ttz\" (UniqueName: \"kubernetes.io/projected/86bfd02c-d226-48e8-bace-e961440bdb43-kube-api-access-x7ttz\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-grqfl\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.576058 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-grqfl\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.580171 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-grqfl\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.580184 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-grqfl\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.580472 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-grqfl\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.580643 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-grqfl\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.582511 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-grqfl\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.583902 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-grqfl\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.603751 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7ttz\" (UniqueName: \"kubernetes.io/projected/86bfd02c-d226-48e8-bace-e961440bdb43-kube-api-access-x7ttz\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-grqfl\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:39:19 crc kubenswrapper[4824]: I1209 10:39:19.658098 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:39:20 crc kubenswrapper[4824]: I1209 10:39:20.420371 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl"] Dec 09 10:39:21 crc kubenswrapper[4824]: I1209 10:39:21.203036 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" event={"ID":"86bfd02c-d226-48e8-bace-e961440bdb43","Type":"ContainerStarted","Data":"23e542160cbe63be2ed74333193c34d397a9fa3e4b1d228dfe4ef1d3616ff285"} Dec 09 10:39:21 crc kubenswrapper[4824]: I1209 10:39:21.203325 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" event={"ID":"86bfd02c-d226-48e8-bace-e961440bdb43","Type":"ContainerStarted","Data":"00a5d71907f5e57ed995fa400f0ddd2a2a3338da4e63001b6d3993b5f48d457d"} Dec 09 10:39:21 crc kubenswrapper[4824]: I1209 10:39:21.226206 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" podStartSLOduration=1.748730003 podStartE2EDuration="2.226160598s" podCreationTimestamp="2025-12-09 10:39:19 +0000 UTC" firstStartedPulling="2025-12-09 10:39:20.429111266 +0000 UTC m=+3116.763615923" lastFinishedPulling="2025-12-09 10:39:20.906541851 +0000 UTC m=+3117.241046518" observedRunningTime="2025-12-09 10:39:21.221047258 +0000 UTC m=+3117.555551925" watchObservedRunningTime="2025-12-09 10:39:21.226160598 +0000 UTC m=+3117.560665275" Dec 09 10:39:32 crc kubenswrapper[4824]: I1209 10:39:32.861194 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:39:32 crc kubenswrapper[4824]: I1209 10:39:32.861729 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:39:32 crc kubenswrapper[4824]: I1209 10:39:32.861774 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 10:39:32 crc kubenswrapper[4824]: I1209 10:39:32.862721 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb"} pod="openshift-machine-config-operator/machine-config-daemon-dth8x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 10:39:32 crc kubenswrapper[4824]: I1209 10:39:32.862804 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" containerID="cri-o://ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" gracePeriod=600 Dec 09 10:39:33 crc kubenswrapper[4824]: E1209 10:39:33.013076 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:39:33 crc kubenswrapper[4824]: I1209 10:39:33.386858 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" exitCode=0 Dec 09 10:39:33 crc kubenswrapper[4824]: I1209 10:39:33.386907 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerDied","Data":"ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb"} Dec 09 10:39:33 crc kubenswrapper[4824]: I1209 10:39:33.386942 4824 scope.go:117] "RemoveContainer" containerID="b963f4bb52756cf9a0c69e35abd4f8d2f8fba82a3075a961f196fd16bdeaec5e" Dec 09 10:39:33 crc kubenswrapper[4824]: I1209 10:39:33.388361 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:39:33 crc kubenswrapper[4824]: E1209 10:39:33.390527 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:39:46 crc kubenswrapper[4824]: I1209 10:39:46.911347 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:39:46 crc kubenswrapper[4824]: E1209 10:39:46.912307 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:39:59 crc kubenswrapper[4824]: I1209 10:39:59.911957 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:39:59 crc kubenswrapper[4824]: E1209 10:39:59.913161 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:40:12 crc kubenswrapper[4824]: I1209 10:40:12.911431 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:40:12 crc kubenswrapper[4824]: E1209 10:40:12.912600 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:40:23 crc kubenswrapper[4824]: I1209 10:40:23.230218 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-t7kkn"] Dec 09 10:40:23 crc kubenswrapper[4824]: I1209 10:40:23.233882 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7kkn" Dec 09 10:40:23 crc kubenswrapper[4824]: I1209 10:40:23.243120 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7kkn"] Dec 09 10:40:23 crc kubenswrapper[4824]: I1209 10:40:23.405369 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjdb9\" (UniqueName: \"kubernetes.io/projected/f3703080-990f-4339-aca1-653914b3c23d-kube-api-access-mjdb9\") pod \"redhat-marketplace-t7kkn\" (UID: \"f3703080-990f-4339-aca1-653914b3c23d\") " pod="openshift-marketplace/redhat-marketplace-t7kkn" Dec 09 10:40:23 crc kubenswrapper[4824]: I1209 10:40:23.405579 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3703080-990f-4339-aca1-653914b3c23d-catalog-content\") pod \"redhat-marketplace-t7kkn\" (UID: \"f3703080-990f-4339-aca1-653914b3c23d\") " pod="openshift-marketplace/redhat-marketplace-t7kkn" Dec 09 10:40:23 crc kubenswrapper[4824]: I1209 10:40:23.405663 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3703080-990f-4339-aca1-653914b3c23d-utilities\") pod \"redhat-marketplace-t7kkn\" (UID: \"f3703080-990f-4339-aca1-653914b3c23d\") " pod="openshift-marketplace/redhat-marketplace-t7kkn" Dec 09 10:40:23 crc kubenswrapper[4824]: I1209 10:40:23.508618 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3703080-990f-4339-aca1-653914b3c23d-utilities\") pod \"redhat-marketplace-t7kkn\" (UID: \"f3703080-990f-4339-aca1-653914b3c23d\") " pod="openshift-marketplace/redhat-marketplace-t7kkn" Dec 09 10:40:23 crc kubenswrapper[4824]: I1209 10:40:23.508809 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjdb9\" (UniqueName: \"kubernetes.io/projected/f3703080-990f-4339-aca1-653914b3c23d-kube-api-access-mjdb9\") pod \"redhat-marketplace-t7kkn\" (UID: \"f3703080-990f-4339-aca1-653914b3c23d\") " pod="openshift-marketplace/redhat-marketplace-t7kkn" Dec 09 10:40:23 crc kubenswrapper[4824]: I1209 10:40:23.509086 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3703080-990f-4339-aca1-653914b3c23d-catalog-content\") pod \"redhat-marketplace-t7kkn\" (UID: \"f3703080-990f-4339-aca1-653914b3c23d\") " pod="openshift-marketplace/redhat-marketplace-t7kkn" Dec 09 10:40:23 crc kubenswrapper[4824]: I1209 10:40:23.509201 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3703080-990f-4339-aca1-653914b3c23d-utilities\") pod \"redhat-marketplace-t7kkn\" (UID: \"f3703080-990f-4339-aca1-653914b3c23d\") " pod="openshift-marketplace/redhat-marketplace-t7kkn" Dec 09 10:40:23 crc kubenswrapper[4824]: I1209 10:40:23.509642 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3703080-990f-4339-aca1-653914b3c23d-catalog-content\") pod \"redhat-marketplace-t7kkn\" (UID: \"f3703080-990f-4339-aca1-653914b3c23d\") " pod="openshift-marketplace/redhat-marketplace-t7kkn" Dec 09 10:40:23 crc kubenswrapper[4824]: I1209 10:40:23.535920 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjdb9\" (UniqueName: \"kubernetes.io/projected/f3703080-990f-4339-aca1-653914b3c23d-kube-api-access-mjdb9\") pod \"redhat-marketplace-t7kkn\" (UID: \"f3703080-990f-4339-aca1-653914b3c23d\") " pod="openshift-marketplace/redhat-marketplace-t7kkn" Dec 09 10:40:23 crc kubenswrapper[4824]: I1209 10:40:23.596032 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7kkn" Dec 09 10:40:24 crc kubenswrapper[4824]: I1209 10:40:24.125950 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7kkn"] Dec 09 10:40:25 crc kubenswrapper[4824]: I1209 10:40:25.122393 4824 generic.go:334] "Generic (PLEG): container finished" podID="f3703080-990f-4339-aca1-653914b3c23d" containerID="b71896b31742fc3179a0eef08f854e4a1f9e86c4f735316e333f04433cb4d6c1" exitCode=0 Dec 09 10:40:25 crc kubenswrapper[4824]: I1209 10:40:25.122579 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7kkn" event={"ID":"f3703080-990f-4339-aca1-653914b3c23d","Type":"ContainerDied","Data":"b71896b31742fc3179a0eef08f854e4a1f9e86c4f735316e333f04433cb4d6c1"} Dec 09 10:40:25 crc kubenswrapper[4824]: I1209 10:40:25.122726 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7kkn" event={"ID":"f3703080-990f-4339-aca1-653914b3c23d","Type":"ContainerStarted","Data":"c248790b5a4f835da115c72fc976349da97a098d8ac6265bcad5b09ab6121a41"} Dec 09 10:40:25 crc kubenswrapper[4824]: I1209 10:40:25.125159 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 10:40:25 crc kubenswrapper[4824]: I1209 10:40:25.589408 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wvpj9"] Dec 09 10:40:25 crc kubenswrapper[4824]: I1209 10:40:25.596609 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wvpj9" Dec 09 10:40:25 crc kubenswrapper[4824]: I1209 10:40:25.606098 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wvpj9"] Dec 09 10:40:25 crc kubenswrapper[4824]: I1209 10:40:25.793056 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26165664-9c17-404b-8c94-a6c0dd593449-utilities\") pod \"certified-operators-wvpj9\" (UID: \"26165664-9c17-404b-8c94-a6c0dd593449\") " pod="openshift-marketplace/certified-operators-wvpj9" Dec 09 10:40:25 crc kubenswrapper[4824]: I1209 10:40:25.793409 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26165664-9c17-404b-8c94-a6c0dd593449-catalog-content\") pod \"certified-operators-wvpj9\" (UID: \"26165664-9c17-404b-8c94-a6c0dd593449\") " pod="openshift-marketplace/certified-operators-wvpj9" Dec 09 10:40:25 crc kubenswrapper[4824]: I1209 10:40:25.793744 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wslbs\" (UniqueName: \"kubernetes.io/projected/26165664-9c17-404b-8c94-a6c0dd593449-kube-api-access-wslbs\") pod \"certified-operators-wvpj9\" (UID: \"26165664-9c17-404b-8c94-a6c0dd593449\") " pod="openshift-marketplace/certified-operators-wvpj9" Dec 09 10:40:25 crc kubenswrapper[4824]: I1209 10:40:25.895260 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26165664-9c17-404b-8c94-a6c0dd593449-utilities\") pod \"certified-operators-wvpj9\" (UID: \"26165664-9c17-404b-8c94-a6c0dd593449\") " pod="openshift-marketplace/certified-operators-wvpj9" Dec 09 10:40:25 crc kubenswrapper[4824]: I1209 10:40:25.895303 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26165664-9c17-404b-8c94-a6c0dd593449-catalog-content\") pod \"certified-operators-wvpj9\" (UID: \"26165664-9c17-404b-8c94-a6c0dd593449\") " pod="openshift-marketplace/certified-operators-wvpj9" Dec 09 10:40:25 crc kubenswrapper[4824]: I1209 10:40:25.895486 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wslbs\" (UniqueName: \"kubernetes.io/projected/26165664-9c17-404b-8c94-a6c0dd593449-kube-api-access-wslbs\") pod \"certified-operators-wvpj9\" (UID: \"26165664-9c17-404b-8c94-a6c0dd593449\") " pod="openshift-marketplace/certified-operators-wvpj9" Dec 09 10:40:25 crc kubenswrapper[4824]: I1209 10:40:25.896008 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26165664-9c17-404b-8c94-a6c0dd593449-catalog-content\") pod \"certified-operators-wvpj9\" (UID: \"26165664-9c17-404b-8c94-a6c0dd593449\") " pod="openshift-marketplace/certified-operators-wvpj9" Dec 09 10:40:25 crc kubenswrapper[4824]: I1209 10:40:25.896000 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26165664-9c17-404b-8c94-a6c0dd593449-utilities\") pod \"certified-operators-wvpj9\" (UID: \"26165664-9c17-404b-8c94-a6c0dd593449\") " pod="openshift-marketplace/certified-operators-wvpj9" Dec 09 10:40:25 crc kubenswrapper[4824]: I1209 10:40:25.926934 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wslbs\" (UniqueName: \"kubernetes.io/projected/26165664-9c17-404b-8c94-a6c0dd593449-kube-api-access-wslbs\") pod \"certified-operators-wvpj9\" (UID: \"26165664-9c17-404b-8c94-a6c0dd593449\") " pod="openshift-marketplace/certified-operators-wvpj9" Dec 09 10:40:25 crc kubenswrapper[4824]: I1209 10:40:25.930868 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wvpj9" Dec 09 10:40:26 crc kubenswrapper[4824]: I1209 10:40:26.452963 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wvpj9"] Dec 09 10:40:26 crc kubenswrapper[4824]: W1209 10:40:26.463307 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26165664_9c17_404b_8c94_a6c0dd593449.slice/crio-6907dd718e099aeafcb77dd126462aa2c752bdd1b6e912f4a130c8788f4f8b68 WatchSource:0}: Error finding container 6907dd718e099aeafcb77dd126462aa2c752bdd1b6e912f4a130c8788f4f8b68: Status 404 returned error can't find the container with id 6907dd718e099aeafcb77dd126462aa2c752bdd1b6e912f4a130c8788f4f8b68 Dec 09 10:40:27 crc kubenswrapper[4824]: I1209 10:40:27.156330 4824 generic.go:334] "Generic (PLEG): container finished" podID="26165664-9c17-404b-8c94-a6c0dd593449" containerID="6cb2360c311b81b8165b040ed44482861266637da0c97cfee469ddcb8487d7ab" exitCode=0 Dec 09 10:40:27 crc kubenswrapper[4824]: I1209 10:40:27.156383 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wvpj9" event={"ID":"26165664-9c17-404b-8c94-a6c0dd593449","Type":"ContainerDied","Data":"6cb2360c311b81b8165b040ed44482861266637da0c97cfee469ddcb8487d7ab"} Dec 09 10:40:27 crc kubenswrapper[4824]: I1209 10:40:27.156716 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wvpj9" event={"ID":"26165664-9c17-404b-8c94-a6c0dd593449","Type":"ContainerStarted","Data":"6907dd718e099aeafcb77dd126462aa2c752bdd1b6e912f4a130c8788f4f8b68"} Dec 09 10:40:27 crc kubenswrapper[4824]: I1209 10:40:27.162766 4824 generic.go:334] "Generic (PLEG): container finished" podID="f3703080-990f-4339-aca1-653914b3c23d" containerID="edf2ca3d19c556c108c0eca61fb1e8a86611cb8ba9078ceeb539c1b63d6773e3" exitCode=0 Dec 09 10:40:27 crc kubenswrapper[4824]: I1209 10:40:27.162835 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7kkn" event={"ID":"f3703080-990f-4339-aca1-653914b3c23d","Type":"ContainerDied","Data":"edf2ca3d19c556c108c0eca61fb1e8a86611cb8ba9078ceeb539c1b63d6773e3"} Dec 09 10:40:27 crc kubenswrapper[4824]: I1209 10:40:27.912930 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:40:27 crc kubenswrapper[4824]: E1209 10:40:27.913546 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:40:28 crc kubenswrapper[4824]: I1209 10:40:28.177486 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wvpj9" event={"ID":"26165664-9c17-404b-8c94-a6c0dd593449","Type":"ContainerStarted","Data":"a3d0857f01a7502d4edea4b3072502ea959907bbcbddade245d373d50e0a180f"} Dec 09 10:40:28 crc kubenswrapper[4824]: I1209 10:40:28.180678 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7kkn" event={"ID":"f3703080-990f-4339-aca1-653914b3c23d","Type":"ContainerStarted","Data":"b5fe3ae666b7d224e77eecf8cdcc87c77ec234c6753012613b80fe4d8abab83e"} Dec 09 10:40:28 crc kubenswrapper[4824]: I1209 10:40:28.250722 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-t7kkn" podStartSLOduration=2.756728655 podStartE2EDuration="5.250693453s" podCreationTimestamp="2025-12-09 10:40:23 +0000 UTC" firstStartedPulling="2025-12-09 10:40:25.124738027 +0000 UTC m=+3181.459242694" lastFinishedPulling="2025-12-09 10:40:27.618702825 +0000 UTC m=+3183.953207492" observedRunningTime="2025-12-09 10:40:28.227385642 +0000 UTC m=+3184.561890309" watchObservedRunningTime="2025-12-09 10:40:28.250693453 +0000 UTC m=+3184.585198120" Dec 09 10:40:30 crc kubenswrapper[4824]: I1209 10:40:30.214165 4824 generic.go:334] "Generic (PLEG): container finished" podID="26165664-9c17-404b-8c94-a6c0dd593449" containerID="a3d0857f01a7502d4edea4b3072502ea959907bbcbddade245d373d50e0a180f" exitCode=0 Dec 09 10:40:30 crc kubenswrapper[4824]: I1209 10:40:30.214567 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wvpj9" event={"ID":"26165664-9c17-404b-8c94-a6c0dd593449","Type":"ContainerDied","Data":"a3d0857f01a7502d4edea4b3072502ea959907bbcbddade245d373d50e0a180f"} Dec 09 10:40:31 crc kubenswrapper[4824]: I1209 10:40:31.231345 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wvpj9" event={"ID":"26165664-9c17-404b-8c94-a6c0dd593449","Type":"ContainerStarted","Data":"33dc0a27704baa4b5c66f664f8a920efead9c3fb50433813581c2728bac3378c"} Dec 09 10:40:31 crc kubenswrapper[4824]: I1209 10:40:31.256645 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wvpj9" podStartSLOduration=2.740294603 podStartE2EDuration="6.25660107s" podCreationTimestamp="2025-12-09 10:40:25 +0000 UTC" firstStartedPulling="2025-12-09 10:40:27.162182286 +0000 UTC m=+3183.496686953" lastFinishedPulling="2025-12-09 10:40:30.678488743 +0000 UTC m=+3187.012993420" observedRunningTime="2025-12-09 10:40:31.251205541 +0000 UTC m=+3187.585710208" watchObservedRunningTime="2025-12-09 10:40:31.25660107 +0000 UTC m=+3187.591105737" Dec 09 10:40:33 crc kubenswrapper[4824]: I1209 10:40:33.597577 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-t7kkn" Dec 09 10:40:33 crc kubenswrapper[4824]: I1209 10:40:33.597922 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-t7kkn" Dec 09 10:40:33 crc kubenswrapper[4824]: I1209 10:40:33.665691 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-t7kkn" Dec 09 10:40:34 crc kubenswrapper[4824]: I1209 10:40:34.339634 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-t7kkn" Dec 09 10:40:34 crc kubenswrapper[4824]: I1209 10:40:34.983727 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7kkn"] Dec 09 10:40:35 crc kubenswrapper[4824]: I1209 10:40:35.932041 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wvpj9" Dec 09 10:40:35 crc kubenswrapper[4824]: I1209 10:40:35.932137 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wvpj9" Dec 09 10:40:36 crc kubenswrapper[4824]: I1209 10:40:36.042439 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wvpj9" Dec 09 10:40:36 crc kubenswrapper[4824]: I1209 10:40:36.315234 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-t7kkn" podUID="f3703080-990f-4339-aca1-653914b3c23d" containerName="registry-server" containerID="cri-o://b5fe3ae666b7d224e77eecf8cdcc87c77ec234c6753012613b80fe4d8abab83e" gracePeriod=2 Dec 09 10:40:36 crc kubenswrapper[4824]: I1209 10:40:36.366332 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wvpj9" Dec 09 10:40:36 crc kubenswrapper[4824]: I1209 10:40:36.816952 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7kkn" Dec 09 10:40:36 crc kubenswrapper[4824]: I1209 10:40:36.954867 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3703080-990f-4339-aca1-653914b3c23d-utilities\") pod \"f3703080-990f-4339-aca1-653914b3c23d\" (UID: \"f3703080-990f-4339-aca1-653914b3c23d\") " Dec 09 10:40:36 crc kubenswrapper[4824]: I1209 10:40:36.955039 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjdb9\" (UniqueName: \"kubernetes.io/projected/f3703080-990f-4339-aca1-653914b3c23d-kube-api-access-mjdb9\") pod \"f3703080-990f-4339-aca1-653914b3c23d\" (UID: \"f3703080-990f-4339-aca1-653914b3c23d\") " Dec 09 10:40:36 crc kubenswrapper[4824]: I1209 10:40:36.955406 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3703080-990f-4339-aca1-653914b3c23d-catalog-content\") pod \"f3703080-990f-4339-aca1-653914b3c23d\" (UID: \"f3703080-990f-4339-aca1-653914b3c23d\") " Dec 09 10:40:36 crc kubenswrapper[4824]: I1209 10:40:36.956069 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3703080-990f-4339-aca1-653914b3c23d-utilities" (OuterVolumeSpecName: "utilities") pod "f3703080-990f-4339-aca1-653914b3c23d" (UID: "f3703080-990f-4339-aca1-653914b3c23d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:40:36 crc kubenswrapper[4824]: I1209 10:40:36.966141 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3703080-990f-4339-aca1-653914b3c23d-kube-api-access-mjdb9" (OuterVolumeSpecName: "kube-api-access-mjdb9") pod "f3703080-990f-4339-aca1-653914b3c23d" (UID: "f3703080-990f-4339-aca1-653914b3c23d"). InnerVolumeSpecName "kube-api-access-mjdb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:40:36 crc kubenswrapper[4824]: I1209 10:40:36.978443 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3703080-990f-4339-aca1-653914b3c23d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f3703080-990f-4339-aca1-653914b3c23d" (UID: "f3703080-990f-4339-aca1-653914b3c23d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:40:37 crc kubenswrapper[4824]: I1209 10:40:37.058269 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3703080-990f-4339-aca1-653914b3c23d-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 10:40:37 crc kubenswrapper[4824]: I1209 10:40:37.058314 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjdb9\" (UniqueName: \"kubernetes.io/projected/f3703080-990f-4339-aca1-653914b3c23d-kube-api-access-mjdb9\") on node \"crc\" DevicePath \"\"" Dec 09 10:40:37 crc kubenswrapper[4824]: I1209 10:40:37.058329 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3703080-990f-4339-aca1-653914b3c23d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 10:40:37 crc kubenswrapper[4824]: I1209 10:40:37.327041 4824 generic.go:334] "Generic (PLEG): container finished" podID="f3703080-990f-4339-aca1-653914b3c23d" containerID="b5fe3ae666b7d224e77eecf8cdcc87c77ec234c6753012613b80fe4d8abab83e" exitCode=0 Dec 09 10:40:37 crc kubenswrapper[4824]: I1209 10:40:37.327107 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7kkn" Dec 09 10:40:37 crc kubenswrapper[4824]: I1209 10:40:37.327094 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7kkn" event={"ID":"f3703080-990f-4339-aca1-653914b3c23d","Type":"ContainerDied","Data":"b5fe3ae666b7d224e77eecf8cdcc87c77ec234c6753012613b80fe4d8abab83e"} Dec 09 10:40:37 crc kubenswrapper[4824]: I1209 10:40:37.327246 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7kkn" event={"ID":"f3703080-990f-4339-aca1-653914b3c23d","Type":"ContainerDied","Data":"c248790b5a4f835da115c72fc976349da97a098d8ac6265bcad5b09ab6121a41"} Dec 09 10:40:37 crc kubenswrapper[4824]: I1209 10:40:37.327265 4824 scope.go:117] "RemoveContainer" containerID="b5fe3ae666b7d224e77eecf8cdcc87c77ec234c6753012613b80fe4d8abab83e" Dec 09 10:40:37 crc kubenswrapper[4824]: I1209 10:40:37.352174 4824 scope.go:117] "RemoveContainer" containerID="edf2ca3d19c556c108c0eca61fb1e8a86611cb8ba9078ceeb539c1b63d6773e3" Dec 09 10:40:37 crc kubenswrapper[4824]: I1209 10:40:37.369628 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7kkn"] Dec 09 10:40:37 crc kubenswrapper[4824]: I1209 10:40:37.381195 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7kkn"] Dec 09 10:40:37 crc kubenswrapper[4824]: I1209 10:40:37.397231 4824 scope.go:117] "RemoveContainer" containerID="b71896b31742fc3179a0eef08f854e4a1f9e86c4f735316e333f04433cb4d6c1" Dec 09 10:40:37 crc kubenswrapper[4824]: I1209 10:40:37.450017 4824 scope.go:117] "RemoveContainer" containerID="b5fe3ae666b7d224e77eecf8cdcc87c77ec234c6753012613b80fe4d8abab83e" Dec 09 10:40:37 crc kubenswrapper[4824]: E1209 10:40:37.450425 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5fe3ae666b7d224e77eecf8cdcc87c77ec234c6753012613b80fe4d8abab83e\": container with ID starting with b5fe3ae666b7d224e77eecf8cdcc87c77ec234c6753012613b80fe4d8abab83e not found: ID does not exist" containerID="b5fe3ae666b7d224e77eecf8cdcc87c77ec234c6753012613b80fe4d8abab83e" Dec 09 10:40:37 crc kubenswrapper[4824]: I1209 10:40:37.450475 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5fe3ae666b7d224e77eecf8cdcc87c77ec234c6753012613b80fe4d8abab83e"} err="failed to get container status \"b5fe3ae666b7d224e77eecf8cdcc87c77ec234c6753012613b80fe4d8abab83e\": rpc error: code = NotFound desc = could not find container \"b5fe3ae666b7d224e77eecf8cdcc87c77ec234c6753012613b80fe4d8abab83e\": container with ID starting with b5fe3ae666b7d224e77eecf8cdcc87c77ec234c6753012613b80fe4d8abab83e not found: ID does not exist" Dec 09 10:40:37 crc kubenswrapper[4824]: I1209 10:40:37.450502 4824 scope.go:117] "RemoveContainer" containerID="edf2ca3d19c556c108c0eca61fb1e8a86611cb8ba9078ceeb539c1b63d6773e3" Dec 09 10:40:37 crc kubenswrapper[4824]: E1209 10:40:37.450713 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edf2ca3d19c556c108c0eca61fb1e8a86611cb8ba9078ceeb539c1b63d6773e3\": container with ID starting with edf2ca3d19c556c108c0eca61fb1e8a86611cb8ba9078ceeb539c1b63d6773e3 not found: ID does not exist" containerID="edf2ca3d19c556c108c0eca61fb1e8a86611cb8ba9078ceeb539c1b63d6773e3" Dec 09 10:40:37 crc kubenswrapper[4824]: I1209 10:40:37.450748 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edf2ca3d19c556c108c0eca61fb1e8a86611cb8ba9078ceeb539c1b63d6773e3"} err="failed to get container status \"edf2ca3d19c556c108c0eca61fb1e8a86611cb8ba9078ceeb539c1b63d6773e3\": rpc error: code = NotFound desc = could not find container \"edf2ca3d19c556c108c0eca61fb1e8a86611cb8ba9078ceeb539c1b63d6773e3\": container with ID starting with edf2ca3d19c556c108c0eca61fb1e8a86611cb8ba9078ceeb539c1b63d6773e3 not found: ID does not exist" Dec 09 10:40:37 crc kubenswrapper[4824]: I1209 10:40:37.450766 4824 scope.go:117] "RemoveContainer" containerID="b71896b31742fc3179a0eef08f854e4a1f9e86c4f735316e333f04433cb4d6c1" Dec 09 10:40:37 crc kubenswrapper[4824]: E1209 10:40:37.451268 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b71896b31742fc3179a0eef08f854e4a1f9e86c4f735316e333f04433cb4d6c1\": container with ID starting with b71896b31742fc3179a0eef08f854e4a1f9e86c4f735316e333f04433cb4d6c1 not found: ID does not exist" containerID="b71896b31742fc3179a0eef08f854e4a1f9e86c4f735316e333f04433cb4d6c1" Dec 09 10:40:37 crc kubenswrapper[4824]: I1209 10:40:37.451307 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b71896b31742fc3179a0eef08f854e4a1f9e86c4f735316e333f04433cb4d6c1"} err="failed to get container status \"b71896b31742fc3179a0eef08f854e4a1f9e86c4f735316e333f04433cb4d6c1\": rpc error: code = NotFound desc = could not find container \"b71896b31742fc3179a0eef08f854e4a1f9e86c4f735316e333f04433cb4d6c1\": container with ID starting with b71896b31742fc3179a0eef08f854e4a1f9e86c4f735316e333f04433cb4d6c1 not found: ID does not exist" Dec 09 10:40:37 crc kubenswrapper[4824]: I1209 10:40:37.902803 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wvpj9"] Dec 09 10:40:37 crc kubenswrapper[4824]: I1209 10:40:37.927873 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3703080-990f-4339-aca1-653914b3c23d" path="/var/lib/kubelet/pods/f3703080-990f-4339-aca1-653914b3c23d/volumes" Dec 09 10:40:38 crc kubenswrapper[4824]: I1209 10:40:38.343435 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wvpj9" podUID="26165664-9c17-404b-8c94-a6c0dd593449" containerName="registry-server" containerID="cri-o://33dc0a27704baa4b5c66f664f8a920efead9c3fb50433813581c2728bac3378c" gracePeriod=2 Dec 09 10:40:38 crc kubenswrapper[4824]: I1209 10:40:38.829693 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wvpj9" Dec 09 10:40:38 crc kubenswrapper[4824]: I1209 10:40:38.906991 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wslbs\" (UniqueName: \"kubernetes.io/projected/26165664-9c17-404b-8c94-a6c0dd593449-kube-api-access-wslbs\") pod \"26165664-9c17-404b-8c94-a6c0dd593449\" (UID: \"26165664-9c17-404b-8c94-a6c0dd593449\") " Dec 09 10:40:38 crc kubenswrapper[4824]: I1209 10:40:38.907192 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26165664-9c17-404b-8c94-a6c0dd593449-catalog-content\") pod \"26165664-9c17-404b-8c94-a6c0dd593449\" (UID: \"26165664-9c17-404b-8c94-a6c0dd593449\") " Dec 09 10:40:38 crc kubenswrapper[4824]: I1209 10:40:38.907338 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26165664-9c17-404b-8c94-a6c0dd593449-utilities\") pod \"26165664-9c17-404b-8c94-a6c0dd593449\" (UID: \"26165664-9c17-404b-8c94-a6c0dd593449\") " Dec 09 10:40:38 crc kubenswrapper[4824]: I1209 10:40:38.908294 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26165664-9c17-404b-8c94-a6c0dd593449-utilities" (OuterVolumeSpecName: "utilities") pod "26165664-9c17-404b-8c94-a6c0dd593449" (UID: "26165664-9c17-404b-8c94-a6c0dd593449"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:40:38 crc kubenswrapper[4824]: I1209 10:40:38.913852 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26165664-9c17-404b-8c94-a6c0dd593449-kube-api-access-wslbs" (OuterVolumeSpecName: "kube-api-access-wslbs") pod "26165664-9c17-404b-8c94-a6c0dd593449" (UID: "26165664-9c17-404b-8c94-a6c0dd593449"). InnerVolumeSpecName "kube-api-access-wslbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:40:38 crc kubenswrapper[4824]: I1209 10:40:38.960861 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26165664-9c17-404b-8c94-a6c0dd593449-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "26165664-9c17-404b-8c94-a6c0dd593449" (UID: "26165664-9c17-404b-8c94-a6c0dd593449"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:40:39 crc kubenswrapper[4824]: I1209 10:40:39.010211 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wslbs\" (UniqueName: \"kubernetes.io/projected/26165664-9c17-404b-8c94-a6c0dd593449-kube-api-access-wslbs\") on node \"crc\" DevicePath \"\"" Dec 09 10:40:39 crc kubenswrapper[4824]: I1209 10:40:39.010253 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26165664-9c17-404b-8c94-a6c0dd593449-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 10:40:39 crc kubenswrapper[4824]: I1209 10:40:39.010262 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26165664-9c17-404b-8c94-a6c0dd593449-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 10:40:39 crc kubenswrapper[4824]: I1209 10:40:39.357501 4824 generic.go:334] "Generic (PLEG): container finished" podID="26165664-9c17-404b-8c94-a6c0dd593449" containerID="33dc0a27704baa4b5c66f664f8a920efead9c3fb50433813581c2728bac3378c" exitCode=0 Dec 09 10:40:39 crc kubenswrapper[4824]: I1209 10:40:39.357545 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wvpj9" event={"ID":"26165664-9c17-404b-8c94-a6c0dd593449","Type":"ContainerDied","Data":"33dc0a27704baa4b5c66f664f8a920efead9c3fb50433813581c2728bac3378c"} Dec 09 10:40:39 crc kubenswrapper[4824]: I1209 10:40:39.357580 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wvpj9" event={"ID":"26165664-9c17-404b-8c94-a6c0dd593449","Type":"ContainerDied","Data":"6907dd718e099aeafcb77dd126462aa2c752bdd1b6e912f4a130c8788f4f8b68"} Dec 09 10:40:39 crc kubenswrapper[4824]: I1209 10:40:39.357597 4824 scope.go:117] "RemoveContainer" containerID="33dc0a27704baa4b5c66f664f8a920efead9c3fb50433813581c2728bac3378c" Dec 09 10:40:39 crc kubenswrapper[4824]: I1209 10:40:39.357552 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wvpj9" Dec 09 10:40:39 crc kubenswrapper[4824]: I1209 10:40:39.400579 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wvpj9"] Dec 09 10:40:39 crc kubenswrapper[4824]: I1209 10:40:39.401527 4824 scope.go:117] "RemoveContainer" containerID="a3d0857f01a7502d4edea4b3072502ea959907bbcbddade245d373d50e0a180f" Dec 09 10:40:39 crc kubenswrapper[4824]: I1209 10:40:39.412803 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wvpj9"] Dec 09 10:40:39 crc kubenswrapper[4824]: I1209 10:40:39.447686 4824 scope.go:117] "RemoveContainer" containerID="6cb2360c311b81b8165b040ed44482861266637da0c97cfee469ddcb8487d7ab" Dec 09 10:40:39 crc kubenswrapper[4824]: I1209 10:40:39.485496 4824 scope.go:117] "RemoveContainer" containerID="33dc0a27704baa4b5c66f664f8a920efead9c3fb50433813581c2728bac3378c" Dec 09 10:40:39 crc kubenswrapper[4824]: E1209 10:40:39.485940 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33dc0a27704baa4b5c66f664f8a920efead9c3fb50433813581c2728bac3378c\": container with ID starting with 33dc0a27704baa4b5c66f664f8a920efead9c3fb50433813581c2728bac3378c not found: ID does not exist" containerID="33dc0a27704baa4b5c66f664f8a920efead9c3fb50433813581c2728bac3378c" Dec 09 10:40:39 crc kubenswrapper[4824]: I1209 10:40:39.485981 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33dc0a27704baa4b5c66f664f8a920efead9c3fb50433813581c2728bac3378c"} err="failed to get container status \"33dc0a27704baa4b5c66f664f8a920efead9c3fb50433813581c2728bac3378c\": rpc error: code = NotFound desc = could not find container \"33dc0a27704baa4b5c66f664f8a920efead9c3fb50433813581c2728bac3378c\": container with ID starting with 33dc0a27704baa4b5c66f664f8a920efead9c3fb50433813581c2728bac3378c not found: ID does not exist" Dec 09 10:40:39 crc kubenswrapper[4824]: I1209 10:40:39.486010 4824 scope.go:117] "RemoveContainer" containerID="a3d0857f01a7502d4edea4b3072502ea959907bbcbddade245d373d50e0a180f" Dec 09 10:40:39 crc kubenswrapper[4824]: E1209 10:40:39.486385 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3d0857f01a7502d4edea4b3072502ea959907bbcbddade245d373d50e0a180f\": container with ID starting with a3d0857f01a7502d4edea4b3072502ea959907bbcbddade245d373d50e0a180f not found: ID does not exist" containerID="a3d0857f01a7502d4edea4b3072502ea959907bbcbddade245d373d50e0a180f" Dec 09 10:40:39 crc kubenswrapper[4824]: I1209 10:40:39.486403 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3d0857f01a7502d4edea4b3072502ea959907bbcbddade245d373d50e0a180f"} err="failed to get container status \"a3d0857f01a7502d4edea4b3072502ea959907bbcbddade245d373d50e0a180f\": rpc error: code = NotFound desc = could not find container \"a3d0857f01a7502d4edea4b3072502ea959907bbcbddade245d373d50e0a180f\": container with ID starting with a3d0857f01a7502d4edea4b3072502ea959907bbcbddade245d373d50e0a180f not found: ID does not exist" Dec 09 10:40:39 crc kubenswrapper[4824]: I1209 10:40:39.486417 4824 scope.go:117] "RemoveContainer" containerID="6cb2360c311b81b8165b040ed44482861266637da0c97cfee469ddcb8487d7ab" Dec 09 10:40:39 crc kubenswrapper[4824]: E1209 10:40:39.486658 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cb2360c311b81b8165b040ed44482861266637da0c97cfee469ddcb8487d7ab\": container with ID starting with 6cb2360c311b81b8165b040ed44482861266637da0c97cfee469ddcb8487d7ab not found: ID does not exist" containerID="6cb2360c311b81b8165b040ed44482861266637da0c97cfee469ddcb8487d7ab" Dec 09 10:40:39 crc kubenswrapper[4824]: I1209 10:40:39.486687 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cb2360c311b81b8165b040ed44482861266637da0c97cfee469ddcb8487d7ab"} err="failed to get container status \"6cb2360c311b81b8165b040ed44482861266637da0c97cfee469ddcb8487d7ab\": rpc error: code = NotFound desc = could not find container \"6cb2360c311b81b8165b040ed44482861266637da0c97cfee469ddcb8487d7ab\": container with ID starting with 6cb2360c311b81b8165b040ed44482861266637da0c97cfee469ddcb8487d7ab not found: ID does not exist" Dec 09 10:40:39 crc kubenswrapper[4824]: I1209 10:40:39.924900 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26165664-9c17-404b-8c94-a6c0dd593449" path="/var/lib/kubelet/pods/26165664-9c17-404b-8c94-a6c0dd593449/volumes" Dec 09 10:40:42 crc kubenswrapper[4824]: I1209 10:40:42.911335 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:40:42 crc kubenswrapper[4824]: E1209 10:40:42.912255 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:40:57 crc kubenswrapper[4824]: I1209 10:40:57.911267 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:40:57 crc kubenswrapper[4824]: E1209 10:40:57.912173 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:41:08 crc kubenswrapper[4824]: I1209 10:41:08.911447 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:41:08 crc kubenswrapper[4824]: E1209 10:41:08.912155 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:41:21 crc kubenswrapper[4824]: I1209 10:41:21.910485 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:41:21 crc kubenswrapper[4824]: E1209 10:41:21.911867 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:41:35 crc kubenswrapper[4824]: I1209 10:41:35.910834 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:41:35 crc kubenswrapper[4824]: E1209 10:41:35.911663 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:41:46 crc kubenswrapper[4824]: I1209 10:41:46.630835 4824 generic.go:334] "Generic (PLEG): container finished" podID="86bfd02c-d226-48e8-bace-e961440bdb43" containerID="23e542160cbe63be2ed74333193c34d397a9fa3e4b1d228dfe4ef1d3616ff285" exitCode=0 Dec 09 10:41:46 crc kubenswrapper[4824]: I1209 10:41:46.630956 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" event={"ID":"86bfd02c-d226-48e8-bace-e961440bdb43","Type":"ContainerDied","Data":"23e542160cbe63be2ed74333193c34d397a9fa3e4b1d228dfe4ef1d3616ff285"} Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.182730 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.386034 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-ceilometer-compute-config-data-0\") pod \"86bfd02c-d226-48e8-bace-e961440bdb43\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.386454 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-telemetry-combined-ca-bundle\") pod \"86bfd02c-d226-48e8-bace-e961440bdb43\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.386535 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-ssh-key\") pod \"86bfd02c-d226-48e8-bace-e961440bdb43\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.386740 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7ttz\" (UniqueName: \"kubernetes.io/projected/86bfd02c-d226-48e8-bace-e961440bdb43-kube-api-access-x7ttz\") pod \"86bfd02c-d226-48e8-bace-e961440bdb43\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.386903 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-ceilometer-compute-config-data-2\") pod \"86bfd02c-d226-48e8-bace-e961440bdb43\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.386982 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-ceilometer-compute-config-data-1\") pod \"86bfd02c-d226-48e8-bace-e961440bdb43\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.387016 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-inventory\") pod \"86bfd02c-d226-48e8-bace-e961440bdb43\" (UID: \"86bfd02c-d226-48e8-bace-e961440bdb43\") " Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.393384 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "86bfd02c-d226-48e8-bace-e961440bdb43" (UID: "86bfd02c-d226-48e8-bace-e961440bdb43"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.394084 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86bfd02c-d226-48e8-bace-e961440bdb43-kube-api-access-x7ttz" (OuterVolumeSpecName: "kube-api-access-x7ttz") pod "86bfd02c-d226-48e8-bace-e961440bdb43" (UID: "86bfd02c-d226-48e8-bace-e961440bdb43"). InnerVolumeSpecName "kube-api-access-x7ttz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.423572 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "86bfd02c-d226-48e8-bace-e961440bdb43" (UID: "86bfd02c-d226-48e8-bace-e961440bdb43"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.439672 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "86bfd02c-d226-48e8-bace-e961440bdb43" (UID: "86bfd02c-d226-48e8-bace-e961440bdb43"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.440138 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-inventory" (OuterVolumeSpecName: "inventory") pod "86bfd02c-d226-48e8-bace-e961440bdb43" (UID: "86bfd02c-d226-48e8-bace-e961440bdb43"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.444712 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "86bfd02c-d226-48e8-bace-e961440bdb43" (UID: "86bfd02c-d226-48e8-bace-e961440bdb43"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.449812 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "86bfd02c-d226-48e8-bace-e961440bdb43" (UID: "86bfd02c-d226-48e8-bace-e961440bdb43"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.490021 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.490054 4824 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.490065 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.490075 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7ttz\" (UniqueName: \"kubernetes.io/projected/86bfd02c-d226-48e8-bace-e961440bdb43-kube-api-access-x7ttz\") on node \"crc\" DevicePath \"\"" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.490083 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.490095 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.490111 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86bfd02c-d226-48e8-bace-e961440bdb43-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.656827 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" event={"ID":"86bfd02c-d226-48e8-bace-e961440bdb43","Type":"ContainerDied","Data":"00a5d71907f5e57ed995fa400f0ddd2a2a3338da4e63001b6d3993b5f48d457d"} Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.657213 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00a5d71907f5e57ed995fa400f0ddd2a2a3338da4e63001b6d3993b5f48d457d" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.656877 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-grqfl" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.815592 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26"] Dec 09 10:41:48 crc kubenswrapper[4824]: E1209 10:41:48.816502 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26165664-9c17-404b-8c94-a6c0dd593449" containerName="extract-content" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.816523 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26165664-9c17-404b-8c94-a6c0dd593449" containerName="extract-content" Dec 09 10:41:48 crc kubenswrapper[4824]: E1209 10:41:48.816552 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3703080-990f-4339-aca1-653914b3c23d" containerName="registry-server" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.816559 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3703080-990f-4339-aca1-653914b3c23d" containerName="registry-server" Dec 09 10:41:48 crc kubenswrapper[4824]: E1209 10:41:48.816575 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26165664-9c17-404b-8c94-a6c0dd593449" containerName="extract-utilities" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.816581 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26165664-9c17-404b-8c94-a6c0dd593449" containerName="extract-utilities" Dec 09 10:41:48 crc kubenswrapper[4824]: E1209 10:41:48.816594 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26165664-9c17-404b-8c94-a6c0dd593449" containerName="registry-server" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.816602 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26165664-9c17-404b-8c94-a6c0dd593449" containerName="registry-server" Dec 09 10:41:48 crc kubenswrapper[4824]: E1209 10:41:48.816634 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3703080-990f-4339-aca1-653914b3c23d" containerName="extract-utilities" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.816643 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3703080-990f-4339-aca1-653914b3c23d" containerName="extract-utilities" Dec 09 10:41:48 crc kubenswrapper[4824]: E1209 10:41:48.816670 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3703080-990f-4339-aca1-653914b3c23d" containerName="extract-content" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.816676 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3703080-990f-4339-aca1-653914b3c23d" containerName="extract-content" Dec 09 10:41:48 crc kubenswrapper[4824]: E1209 10:41:48.816686 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86bfd02c-d226-48e8-bace-e961440bdb43" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.816693 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="86bfd02c-d226-48e8-bace-e961440bdb43" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.817211 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="86bfd02c-d226-48e8-bace-e961440bdb43" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.817240 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3703080-990f-4339-aca1-653914b3c23d" containerName="registry-server" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.817266 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26165664-9c17-404b-8c94-a6c0dd593449" containerName="registry-server" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.819100 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.821994 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.822223 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5nzld" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.822294 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.823410 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-ipmi-config-data" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.823656 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.831158 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26"] Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.912363 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.912426 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.912453 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjmvp\" (UniqueName: \"kubernetes.io/projected/c62caaf5-b010-47ee-9987-50cd9b418a28-kube-api-access-rjmvp\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.912505 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.912589 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.912662 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:41:48 crc kubenswrapper[4824]: I1209 10:41:48.912699 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:41:49 crc kubenswrapper[4824]: I1209 10:41:49.015370 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:41:49 crc kubenswrapper[4824]: I1209 10:41:49.015436 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:41:49 crc kubenswrapper[4824]: I1209 10:41:49.015518 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:41:49 crc kubenswrapper[4824]: I1209 10:41:49.015558 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:41:49 crc kubenswrapper[4824]: I1209 10:41:49.015580 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjmvp\" (UniqueName: \"kubernetes.io/projected/c62caaf5-b010-47ee-9987-50cd9b418a28-kube-api-access-rjmvp\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:41:49 crc kubenswrapper[4824]: I1209 10:41:49.015613 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:41:49 crc kubenswrapper[4824]: I1209 10:41:49.015687 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:41:49 crc kubenswrapper[4824]: I1209 10:41:49.019588 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:41:49 crc kubenswrapper[4824]: I1209 10:41:49.019695 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:41:49 crc kubenswrapper[4824]: I1209 10:41:49.019738 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:41:49 crc kubenswrapper[4824]: I1209 10:41:49.020504 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:41:49 crc kubenswrapper[4824]: I1209 10:41:49.020913 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:41:49 crc kubenswrapper[4824]: I1209 10:41:49.035953 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:41:49 crc kubenswrapper[4824]: I1209 10:41:49.042901 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjmvp\" (UniqueName: \"kubernetes.io/projected/c62caaf5-b010-47ee-9987-50cd9b418a28-kube-api-access-rjmvp\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:41:49 crc kubenswrapper[4824]: I1209 10:41:49.143349 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:41:49 crc kubenswrapper[4824]: I1209 10:41:49.785363 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26"] Dec 09 10:41:50 crc kubenswrapper[4824]: I1209 10:41:50.680355 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" event={"ID":"c62caaf5-b010-47ee-9987-50cd9b418a28","Type":"ContainerStarted","Data":"9c83175ee042fd3c5980243b806d2a20c42a8b674a0d212aeed9044cf7ed15e8"} Dec 09 10:41:50 crc kubenswrapper[4824]: I1209 10:41:50.680656 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" event={"ID":"c62caaf5-b010-47ee-9987-50cd9b418a28","Type":"ContainerStarted","Data":"6294f1f4998579b1293eb114a4c852657c54177c58e1f01cce9650bffc18ce31"} Dec 09 10:41:50 crc kubenswrapper[4824]: I1209 10:41:50.711294 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" podStartSLOduration=2.178670889 podStartE2EDuration="2.711264646s" podCreationTimestamp="2025-12-09 10:41:48 +0000 UTC" firstStartedPulling="2025-12-09 10:41:49.790847158 +0000 UTC m=+3266.125351825" lastFinishedPulling="2025-12-09 10:41:50.323440915 +0000 UTC m=+3266.657945582" observedRunningTime="2025-12-09 10:41:50.704299227 +0000 UTC m=+3267.038803914" watchObservedRunningTime="2025-12-09 10:41:50.711264646 +0000 UTC m=+3267.045769313" Dec 09 10:41:50 crc kubenswrapper[4824]: I1209 10:41:50.911154 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:41:50 crc kubenswrapper[4824]: E1209 10:41:50.911515 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:42:01 crc kubenswrapper[4824]: I1209 10:42:01.910951 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:42:01 crc kubenswrapper[4824]: E1209 10:42:01.911841 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:42:14 crc kubenswrapper[4824]: I1209 10:42:14.911039 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:42:14 crc kubenswrapper[4824]: E1209 10:42:14.911839 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:42:25 crc kubenswrapper[4824]: I1209 10:42:25.910896 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:42:25 crc kubenswrapper[4824]: E1209 10:42:25.911763 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:42:38 crc kubenswrapper[4824]: I1209 10:42:38.910585 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:42:38 crc kubenswrapper[4824]: E1209 10:42:38.911857 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:42:51 crc kubenswrapper[4824]: I1209 10:42:51.910889 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:42:51 crc kubenswrapper[4824]: E1209 10:42:51.911737 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:43:03 crc kubenswrapper[4824]: I1209 10:43:03.931840 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:43:03 crc kubenswrapper[4824]: E1209 10:43:03.932437 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:43:17 crc kubenswrapper[4824]: I1209 10:43:17.911091 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:43:17 crc kubenswrapper[4824]: E1209 10:43:17.912164 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:43:29 crc kubenswrapper[4824]: I1209 10:43:29.912818 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:43:29 crc kubenswrapper[4824]: E1209 10:43:29.914612 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:43:42 crc kubenswrapper[4824]: I1209 10:43:42.910697 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:43:42 crc kubenswrapper[4824]: E1209 10:43:42.912237 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:43:54 crc kubenswrapper[4824]: I1209 10:43:54.911092 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:43:54 crc kubenswrapper[4824]: E1209 10:43:54.912931 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:44:01 crc kubenswrapper[4824]: I1209 10:44:01.508136 4824 generic.go:334] "Generic (PLEG): container finished" podID="c62caaf5-b010-47ee-9987-50cd9b418a28" containerID="9c83175ee042fd3c5980243b806d2a20c42a8b674a0d212aeed9044cf7ed15e8" exitCode=0 Dec 09 10:44:01 crc kubenswrapper[4824]: I1209 10:44:01.508620 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" event={"ID":"c62caaf5-b010-47ee-9987-50cd9b418a28","Type":"ContainerDied","Data":"9c83175ee042fd3c5980243b806d2a20c42a8b674a0d212aeed9044cf7ed15e8"} Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.099016 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.232052 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-telemetry-power-monitoring-combined-ca-bundle\") pod \"c62caaf5-b010-47ee-9987-50cd9b418a28\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.232112 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-ceilometer-ipmi-config-data-1\") pod \"c62caaf5-b010-47ee-9987-50cd9b418a28\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.232137 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-inventory\") pod \"c62caaf5-b010-47ee-9987-50cd9b418a28\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.232189 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-ceilometer-ipmi-config-data-2\") pod \"c62caaf5-b010-47ee-9987-50cd9b418a28\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.232238 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-ceilometer-ipmi-config-data-0\") pod \"c62caaf5-b010-47ee-9987-50cd9b418a28\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.232257 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjmvp\" (UniqueName: \"kubernetes.io/projected/c62caaf5-b010-47ee-9987-50cd9b418a28-kube-api-access-rjmvp\") pod \"c62caaf5-b010-47ee-9987-50cd9b418a28\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.232376 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-ssh-key\") pod \"c62caaf5-b010-47ee-9987-50cd9b418a28\" (UID: \"c62caaf5-b010-47ee-9987-50cd9b418a28\") " Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.238227 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c62caaf5-b010-47ee-9987-50cd9b418a28-kube-api-access-rjmvp" (OuterVolumeSpecName: "kube-api-access-rjmvp") pod "c62caaf5-b010-47ee-9987-50cd9b418a28" (UID: "c62caaf5-b010-47ee-9987-50cd9b418a28"). InnerVolumeSpecName "kube-api-access-rjmvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.238714 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "c62caaf5-b010-47ee-9987-50cd9b418a28" (UID: "c62caaf5-b010-47ee-9987-50cd9b418a28"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.267989 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-ceilometer-ipmi-config-data-1" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-1") pod "c62caaf5-b010-47ee-9987-50cd9b418a28" (UID: "c62caaf5-b010-47ee-9987-50cd9b418a28"). InnerVolumeSpecName "ceilometer-ipmi-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.269470 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-inventory" (OuterVolumeSpecName: "inventory") pod "c62caaf5-b010-47ee-9987-50cd9b418a28" (UID: "c62caaf5-b010-47ee-9987-50cd9b418a28"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.270475 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-ceilometer-ipmi-config-data-0" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-0") pod "c62caaf5-b010-47ee-9987-50cd9b418a28" (UID: "c62caaf5-b010-47ee-9987-50cd9b418a28"). InnerVolumeSpecName "ceilometer-ipmi-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.273845 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-ceilometer-ipmi-config-data-2" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-2") pod "c62caaf5-b010-47ee-9987-50cd9b418a28" (UID: "c62caaf5-b010-47ee-9987-50cd9b418a28"). InnerVolumeSpecName "ceilometer-ipmi-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.293670 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c62caaf5-b010-47ee-9987-50cd9b418a28" (UID: "c62caaf5-b010-47ee-9987-50cd9b418a28"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.336378 4824 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.336435 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-ceilometer-ipmi-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.336452 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.336464 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-ceilometer-ipmi-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.336476 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-ceilometer-ipmi-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.336489 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjmvp\" (UniqueName: \"kubernetes.io/projected/c62caaf5-b010-47ee-9987-50cd9b418a28-kube-api-access-rjmvp\") on node \"crc\" DevicePath \"\"" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.336501 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c62caaf5-b010-47ee-9987-50cd9b418a28-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.533104 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" event={"ID":"c62caaf5-b010-47ee-9987-50cd9b418a28","Type":"ContainerDied","Data":"6294f1f4998579b1293eb114a4c852657c54177c58e1f01cce9650bffc18ce31"} Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.533150 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6294f1f4998579b1293eb114a4c852657c54177c58e1f01cce9650bffc18ce31" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.533594 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.682326 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749"] Dec 09 10:44:03 crc kubenswrapper[4824]: E1209 10:44:03.682927 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c62caaf5-b010-47ee-9987-50cd9b418a28" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.682947 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c62caaf5-b010-47ee-9987-50cd9b418a28" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.683221 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c62caaf5-b010-47ee-9987-50cd9b418a28" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.684167 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.687078 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.687097 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"logging-compute-config-data" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.687104 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5nzld" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.687163 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.687180 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 10:44:03 crc kubenswrapper[4824]: I1209 10:44:03.692468 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749"] Dec 09 10:44:04 crc kubenswrapper[4824]: I1209 10:44:03.746671 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-hj749\" (UID: \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749" Dec 09 10:44:04 crc kubenswrapper[4824]: I1209 10:44:03.746746 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-hj749\" (UID: \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749" Dec 09 10:44:04 crc kubenswrapper[4824]: I1209 10:44:03.746805 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-hj749\" (UID: \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749" Dec 09 10:44:04 crc kubenswrapper[4824]: I1209 10:44:03.746911 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-hj749\" (UID: \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749" Dec 09 10:44:04 crc kubenswrapper[4824]: I1209 10:44:03.747082 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hccsg\" (UniqueName: \"kubernetes.io/projected/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-kube-api-access-hccsg\") pod \"logging-edpm-deployment-openstack-edpm-ipam-hj749\" (UID: \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749" Dec 09 10:44:04 crc kubenswrapper[4824]: I1209 10:44:03.849448 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-hj749\" (UID: \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749" Dec 09 10:44:04 crc kubenswrapper[4824]: I1209 10:44:03.849539 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-hj749\" (UID: \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749" Dec 09 10:44:04 crc kubenswrapper[4824]: I1209 10:44:03.849576 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-hj749\" (UID: \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749" Dec 09 10:44:04 crc kubenswrapper[4824]: I1209 10:44:03.849681 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-hj749\" (UID: \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749" Dec 09 10:44:04 crc kubenswrapper[4824]: I1209 10:44:03.849893 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hccsg\" (UniqueName: \"kubernetes.io/projected/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-kube-api-access-hccsg\") pod \"logging-edpm-deployment-openstack-edpm-ipam-hj749\" (UID: \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749" Dec 09 10:44:04 crc kubenswrapper[4824]: I1209 10:44:03.854930 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-hj749\" (UID: \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749" Dec 09 10:44:04 crc kubenswrapper[4824]: I1209 10:44:03.855354 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-hj749\" (UID: \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749" Dec 09 10:44:04 crc kubenswrapper[4824]: I1209 10:44:03.855739 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-hj749\" (UID: \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749" Dec 09 10:44:04 crc kubenswrapper[4824]: I1209 10:44:03.858385 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-hj749\" (UID: \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749" Dec 09 10:44:04 crc kubenswrapper[4824]: I1209 10:44:03.870296 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hccsg\" (UniqueName: \"kubernetes.io/projected/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-kube-api-access-hccsg\") pod \"logging-edpm-deployment-openstack-edpm-ipam-hj749\" (UID: \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749" Dec 09 10:44:04 crc kubenswrapper[4824]: I1209 10:44:04.010646 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749" Dec 09 10:44:05 crc kubenswrapper[4824]: I1209 10:44:05.018608 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749"] Dec 09 10:44:05 crc kubenswrapper[4824]: I1209 10:44:05.587721 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749" event={"ID":"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be","Type":"ContainerStarted","Data":"4fc5a3c2f7bf9b6c2e69a79a2c98c79ec7ec2fb698e2bff363b98225343e3e5d"} Dec 09 10:44:06 crc kubenswrapper[4824]: I1209 10:44:06.600310 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749" event={"ID":"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be","Type":"ContainerStarted","Data":"d513485f24d4790143759544f5b3b49ef1f739cd4f133d2e81fd6ce1a578c523"} Dec 09 10:44:06 crc kubenswrapper[4824]: I1209 10:44:06.616574 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749" podStartSLOduration=3.118523557 podStartE2EDuration="3.616549302s" podCreationTimestamp="2025-12-09 10:44:03 +0000 UTC" firstStartedPulling="2025-12-09 10:44:05.009841372 +0000 UTC m=+3401.344346039" lastFinishedPulling="2025-12-09 10:44:05.507867117 +0000 UTC m=+3401.842371784" observedRunningTime="2025-12-09 10:44:06.614634182 +0000 UTC m=+3402.949138869" watchObservedRunningTime="2025-12-09 10:44:06.616549302 +0000 UTC m=+3402.951053969" Dec 09 10:44:07 crc kubenswrapper[4824]: I1209 10:44:07.911606 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:44:07 crc kubenswrapper[4824]: E1209 10:44:07.912483 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:44:21 crc kubenswrapper[4824]: I1209 10:44:21.911110 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:44:21 crc kubenswrapper[4824]: E1209 10:44:21.911978 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:44:23 crc kubenswrapper[4824]: I1209 10:44:23.174718 4824 generic.go:334] "Generic (PLEG): container finished" podID="edcacc59-d95a-42e3-8bc4-9e8a1a7df2be" containerID="d513485f24d4790143759544f5b3b49ef1f739cd4f133d2e81fd6ce1a578c523" exitCode=0 Dec 09 10:44:23 crc kubenswrapper[4824]: I1209 10:44:23.174861 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749" event={"ID":"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be","Type":"ContainerDied","Data":"d513485f24d4790143759544f5b3b49ef1f739cd4f133d2e81fd6ce1a578c523"} Dec 09 10:44:24 crc kubenswrapper[4824]: I1209 10:44:24.662258 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749" Dec 09 10:44:24 crc kubenswrapper[4824]: I1209 10:44:24.770099 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-inventory\") pod \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\" (UID: \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\") " Dec 09 10:44:24 crc kubenswrapper[4824]: I1209 10:44:24.770314 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-ssh-key\") pod \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\" (UID: \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\") " Dec 09 10:44:24 crc kubenswrapper[4824]: I1209 10:44:24.770441 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-logging-compute-config-data-1\") pod \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\" (UID: \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\") " Dec 09 10:44:24 crc kubenswrapper[4824]: I1209 10:44:24.770462 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-logging-compute-config-data-0\") pod \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\" (UID: \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\") " Dec 09 10:44:24 crc kubenswrapper[4824]: I1209 10:44:24.770587 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hccsg\" (UniqueName: \"kubernetes.io/projected/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-kube-api-access-hccsg\") pod \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\" (UID: \"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be\") " Dec 09 10:44:24 crc kubenswrapper[4824]: I1209 10:44:24.779052 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-kube-api-access-hccsg" (OuterVolumeSpecName: "kube-api-access-hccsg") pod "edcacc59-d95a-42e3-8bc4-9e8a1a7df2be" (UID: "edcacc59-d95a-42e3-8bc4-9e8a1a7df2be"). InnerVolumeSpecName "kube-api-access-hccsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:44:24 crc kubenswrapper[4824]: I1209 10:44:24.807039 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-logging-compute-config-data-0" (OuterVolumeSpecName: "logging-compute-config-data-0") pod "edcacc59-d95a-42e3-8bc4-9e8a1a7df2be" (UID: "edcacc59-d95a-42e3-8bc4-9e8a1a7df2be"). InnerVolumeSpecName "logging-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:44:24 crc kubenswrapper[4824]: I1209 10:44:24.814507 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "edcacc59-d95a-42e3-8bc4-9e8a1a7df2be" (UID: "edcacc59-d95a-42e3-8bc4-9e8a1a7df2be"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:44:24 crc kubenswrapper[4824]: I1209 10:44:24.816769 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-logging-compute-config-data-1" (OuterVolumeSpecName: "logging-compute-config-data-1") pod "edcacc59-d95a-42e3-8bc4-9e8a1a7df2be" (UID: "edcacc59-d95a-42e3-8bc4-9e8a1a7df2be"). InnerVolumeSpecName "logging-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:44:24 crc kubenswrapper[4824]: I1209 10:44:24.837202 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-inventory" (OuterVolumeSpecName: "inventory") pod "edcacc59-d95a-42e3-8bc4-9e8a1a7df2be" (UID: "edcacc59-d95a-42e3-8bc4-9e8a1a7df2be"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:44:24 crc kubenswrapper[4824]: I1209 10:44:24.874609 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 10:44:24 crc kubenswrapper[4824]: I1209 10:44:24.874858 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 10:44:24 crc kubenswrapper[4824]: I1209 10:44:24.874933 4824 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-logging-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 09 10:44:24 crc kubenswrapper[4824]: I1209 10:44:24.874999 4824 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-logging-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 09 10:44:24 crc kubenswrapper[4824]: I1209 10:44:24.875088 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hccsg\" (UniqueName: \"kubernetes.io/projected/edcacc59-d95a-42e3-8bc4-9e8a1a7df2be-kube-api-access-hccsg\") on node \"crc\" DevicePath \"\"" Dec 09 10:44:25 crc kubenswrapper[4824]: I1209 10:44:25.198511 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749" event={"ID":"edcacc59-d95a-42e3-8bc4-9e8a1a7df2be","Type":"ContainerDied","Data":"4fc5a3c2f7bf9b6c2e69a79a2c98c79ec7ec2fb698e2bff363b98225343e3e5d"} Dec 09 10:44:25 crc kubenswrapper[4824]: I1209 10:44:25.198554 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fc5a3c2f7bf9b6c2e69a79a2c98c79ec7ec2fb698e2bff363b98225343e3e5d" Dec 09 10:44:25 crc kubenswrapper[4824]: I1209 10:44:25.198805 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-hj749" Dec 09 10:44:32 crc kubenswrapper[4824]: I1209 10:44:32.911756 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:44:33 crc kubenswrapper[4824]: I1209 10:44:33.303483 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerStarted","Data":"7050a696e11544a6263f8787d7539e141eb5adea02be70809cc2757af296657c"} Dec 09 10:44:48 crc kubenswrapper[4824]: I1209 10:44:48.767720 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wk72l"] Dec 09 10:44:48 crc kubenswrapper[4824]: E1209 10:44:48.770131 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edcacc59-d95a-42e3-8bc4-9e8a1a7df2be" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 09 10:44:48 crc kubenswrapper[4824]: I1209 10:44:48.770434 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="edcacc59-d95a-42e3-8bc4-9e8a1a7df2be" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 09 10:44:48 crc kubenswrapper[4824]: I1209 10:44:48.770880 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="edcacc59-d95a-42e3-8bc4-9e8a1a7df2be" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 09 10:44:48 crc kubenswrapper[4824]: I1209 10:44:48.773228 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wk72l" Dec 09 10:44:48 crc kubenswrapper[4824]: I1209 10:44:48.787907 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wk72l"] Dec 09 10:44:48 crc kubenswrapper[4824]: I1209 10:44:48.919741 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be531b94-496d-45d7-9b5e-6532bd4f9890-utilities\") pod \"redhat-operators-wk72l\" (UID: \"be531b94-496d-45d7-9b5e-6532bd4f9890\") " pod="openshift-marketplace/redhat-operators-wk72l" Dec 09 10:44:48 crc kubenswrapper[4824]: I1209 10:44:48.920382 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be531b94-496d-45d7-9b5e-6532bd4f9890-catalog-content\") pod \"redhat-operators-wk72l\" (UID: \"be531b94-496d-45d7-9b5e-6532bd4f9890\") " pod="openshift-marketplace/redhat-operators-wk72l" Dec 09 10:44:48 crc kubenswrapper[4824]: I1209 10:44:48.920574 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qc8b\" (UniqueName: \"kubernetes.io/projected/be531b94-496d-45d7-9b5e-6532bd4f9890-kube-api-access-9qc8b\") pod \"redhat-operators-wk72l\" (UID: \"be531b94-496d-45d7-9b5e-6532bd4f9890\") " pod="openshift-marketplace/redhat-operators-wk72l" Dec 09 10:44:49 crc kubenswrapper[4824]: I1209 10:44:49.023194 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be531b94-496d-45d7-9b5e-6532bd4f9890-catalog-content\") pod \"redhat-operators-wk72l\" (UID: \"be531b94-496d-45d7-9b5e-6532bd4f9890\") " pod="openshift-marketplace/redhat-operators-wk72l" Dec 09 10:44:49 crc kubenswrapper[4824]: I1209 10:44:49.023435 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qc8b\" (UniqueName: \"kubernetes.io/projected/be531b94-496d-45d7-9b5e-6532bd4f9890-kube-api-access-9qc8b\") pod \"redhat-operators-wk72l\" (UID: \"be531b94-496d-45d7-9b5e-6532bd4f9890\") " pod="openshift-marketplace/redhat-operators-wk72l" Dec 09 10:44:49 crc kubenswrapper[4824]: I1209 10:44:49.023682 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be531b94-496d-45d7-9b5e-6532bd4f9890-utilities\") pod \"redhat-operators-wk72l\" (UID: \"be531b94-496d-45d7-9b5e-6532bd4f9890\") " pod="openshift-marketplace/redhat-operators-wk72l" Dec 09 10:44:49 crc kubenswrapper[4824]: I1209 10:44:49.024859 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be531b94-496d-45d7-9b5e-6532bd4f9890-catalog-content\") pod \"redhat-operators-wk72l\" (UID: \"be531b94-496d-45d7-9b5e-6532bd4f9890\") " pod="openshift-marketplace/redhat-operators-wk72l" Dec 09 10:44:49 crc kubenswrapper[4824]: I1209 10:44:49.026574 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be531b94-496d-45d7-9b5e-6532bd4f9890-utilities\") pod \"redhat-operators-wk72l\" (UID: \"be531b94-496d-45d7-9b5e-6532bd4f9890\") " pod="openshift-marketplace/redhat-operators-wk72l" Dec 09 10:44:49 crc kubenswrapper[4824]: I1209 10:44:49.055751 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qc8b\" (UniqueName: \"kubernetes.io/projected/be531b94-496d-45d7-9b5e-6532bd4f9890-kube-api-access-9qc8b\") pod \"redhat-operators-wk72l\" (UID: \"be531b94-496d-45d7-9b5e-6532bd4f9890\") " pod="openshift-marketplace/redhat-operators-wk72l" Dec 09 10:44:49 crc kubenswrapper[4824]: I1209 10:44:49.109161 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wk72l" Dec 09 10:44:49 crc kubenswrapper[4824]: I1209 10:44:49.739449 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wk72l"] Dec 09 10:44:50 crc kubenswrapper[4824]: I1209 10:44:50.508703 4824 generic.go:334] "Generic (PLEG): container finished" podID="be531b94-496d-45d7-9b5e-6532bd4f9890" containerID="9f6b541c1d3e8df40a30f6c785c1fba074b6f3db4808b71565d62c29f578229e" exitCode=0 Dec 09 10:44:50 crc kubenswrapper[4824]: I1209 10:44:50.509008 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wk72l" event={"ID":"be531b94-496d-45d7-9b5e-6532bd4f9890","Type":"ContainerDied","Data":"9f6b541c1d3e8df40a30f6c785c1fba074b6f3db4808b71565d62c29f578229e"} Dec 09 10:44:50 crc kubenswrapper[4824]: I1209 10:44:50.509039 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wk72l" event={"ID":"be531b94-496d-45d7-9b5e-6532bd4f9890","Type":"ContainerStarted","Data":"6b2717a3ec5ed885482dcee97baf865eae8c26f632a8159c532c9d4cd33c0e7a"} Dec 09 10:44:51 crc kubenswrapper[4824]: I1209 10:44:51.523347 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wk72l" event={"ID":"be531b94-496d-45d7-9b5e-6532bd4f9890","Type":"ContainerStarted","Data":"995fa5682ba3e4e72b03173615f379ded0568046a23780ec66d954cc43143534"} Dec 09 10:44:52 crc kubenswrapper[4824]: I1209 10:44:52.357325 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9jnd8"] Dec 09 10:44:52 crc kubenswrapper[4824]: I1209 10:44:52.359833 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9jnd8" Dec 09 10:44:52 crc kubenswrapper[4824]: I1209 10:44:52.372109 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9jnd8"] Dec 09 10:44:52 crc kubenswrapper[4824]: I1209 10:44:52.430286 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klhl9\" (UniqueName: \"kubernetes.io/projected/cc82131a-c612-48b6-be27-495cebd6296b-kube-api-access-klhl9\") pod \"community-operators-9jnd8\" (UID: \"cc82131a-c612-48b6-be27-495cebd6296b\") " pod="openshift-marketplace/community-operators-9jnd8" Dec 09 10:44:52 crc kubenswrapper[4824]: I1209 10:44:52.430382 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc82131a-c612-48b6-be27-495cebd6296b-catalog-content\") pod \"community-operators-9jnd8\" (UID: \"cc82131a-c612-48b6-be27-495cebd6296b\") " pod="openshift-marketplace/community-operators-9jnd8" Dec 09 10:44:52 crc kubenswrapper[4824]: I1209 10:44:52.430485 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc82131a-c612-48b6-be27-495cebd6296b-utilities\") pod \"community-operators-9jnd8\" (UID: \"cc82131a-c612-48b6-be27-495cebd6296b\") " pod="openshift-marketplace/community-operators-9jnd8" Dec 09 10:44:52 crc kubenswrapper[4824]: I1209 10:44:52.532332 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klhl9\" (UniqueName: \"kubernetes.io/projected/cc82131a-c612-48b6-be27-495cebd6296b-kube-api-access-klhl9\") pod \"community-operators-9jnd8\" (UID: \"cc82131a-c612-48b6-be27-495cebd6296b\") " pod="openshift-marketplace/community-operators-9jnd8" Dec 09 10:44:52 crc kubenswrapper[4824]: I1209 10:44:52.532655 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc82131a-c612-48b6-be27-495cebd6296b-catalog-content\") pod \"community-operators-9jnd8\" (UID: \"cc82131a-c612-48b6-be27-495cebd6296b\") " pod="openshift-marketplace/community-operators-9jnd8" Dec 09 10:44:52 crc kubenswrapper[4824]: I1209 10:44:52.532709 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc82131a-c612-48b6-be27-495cebd6296b-utilities\") pod \"community-operators-9jnd8\" (UID: \"cc82131a-c612-48b6-be27-495cebd6296b\") " pod="openshift-marketplace/community-operators-9jnd8" Dec 09 10:44:52 crc kubenswrapper[4824]: I1209 10:44:52.533263 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc82131a-c612-48b6-be27-495cebd6296b-utilities\") pod \"community-operators-9jnd8\" (UID: \"cc82131a-c612-48b6-be27-495cebd6296b\") " pod="openshift-marketplace/community-operators-9jnd8" Dec 09 10:44:52 crc kubenswrapper[4824]: I1209 10:44:52.533453 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc82131a-c612-48b6-be27-495cebd6296b-catalog-content\") pod \"community-operators-9jnd8\" (UID: \"cc82131a-c612-48b6-be27-495cebd6296b\") " pod="openshift-marketplace/community-operators-9jnd8" Dec 09 10:44:52 crc kubenswrapper[4824]: I1209 10:44:52.723398 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klhl9\" (UniqueName: \"kubernetes.io/projected/cc82131a-c612-48b6-be27-495cebd6296b-kube-api-access-klhl9\") pod \"community-operators-9jnd8\" (UID: \"cc82131a-c612-48b6-be27-495cebd6296b\") " pod="openshift-marketplace/community-operators-9jnd8" Dec 09 10:44:52 crc kubenswrapper[4824]: I1209 10:44:52.993606 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9jnd8" Dec 09 10:44:53 crc kubenswrapper[4824]: I1209 10:44:53.548530 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9jnd8"] Dec 09 10:44:54 crc kubenswrapper[4824]: I1209 10:44:54.564267 4824 generic.go:334] "Generic (PLEG): container finished" podID="be531b94-496d-45d7-9b5e-6532bd4f9890" containerID="995fa5682ba3e4e72b03173615f379ded0568046a23780ec66d954cc43143534" exitCode=0 Dec 09 10:44:54 crc kubenswrapper[4824]: I1209 10:44:54.564321 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wk72l" event={"ID":"be531b94-496d-45d7-9b5e-6532bd4f9890","Type":"ContainerDied","Data":"995fa5682ba3e4e72b03173615f379ded0568046a23780ec66d954cc43143534"} Dec 09 10:44:54 crc kubenswrapper[4824]: I1209 10:44:54.567535 4824 generic.go:334] "Generic (PLEG): container finished" podID="cc82131a-c612-48b6-be27-495cebd6296b" containerID="3620580e79bb5e72eef0a1cec5838c5aed128251bcf9ca3c97fbeebbbd4d8420" exitCode=0 Dec 09 10:44:54 crc kubenswrapper[4824]: I1209 10:44:54.567725 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9jnd8" event={"ID":"cc82131a-c612-48b6-be27-495cebd6296b","Type":"ContainerDied","Data":"3620580e79bb5e72eef0a1cec5838c5aed128251bcf9ca3c97fbeebbbd4d8420"} Dec 09 10:44:54 crc kubenswrapper[4824]: I1209 10:44:54.567911 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9jnd8" event={"ID":"cc82131a-c612-48b6-be27-495cebd6296b","Type":"ContainerStarted","Data":"f330bf70af3fd4b6c197529c1648f74c8e5d63743d843e9b0bc5c1d75e4ec060"} Dec 09 10:44:55 crc kubenswrapper[4824]: I1209 10:44:55.580072 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wk72l" event={"ID":"be531b94-496d-45d7-9b5e-6532bd4f9890","Type":"ContainerStarted","Data":"271b56f2e4ef0073d0c4f22851310329a0bbc483f52691cd44c45164e60467c2"} Dec 09 10:44:55 crc kubenswrapper[4824]: I1209 10:44:55.588660 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9jnd8" event={"ID":"cc82131a-c612-48b6-be27-495cebd6296b","Type":"ContainerStarted","Data":"f77652e6456ac553187dff866b75f91dea702397ccf685cdc1cb4dc06c532017"} Dec 09 10:44:55 crc kubenswrapper[4824]: I1209 10:44:55.611328 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wk72l" podStartSLOduration=3.144018516 podStartE2EDuration="7.611306871s" podCreationTimestamp="2025-12-09 10:44:48 +0000 UTC" firstStartedPulling="2025-12-09 10:44:50.513104619 +0000 UTC m=+3446.847609286" lastFinishedPulling="2025-12-09 10:44:54.980392974 +0000 UTC m=+3451.314897641" observedRunningTime="2025-12-09 10:44:55.603596892 +0000 UTC m=+3451.938101559" watchObservedRunningTime="2025-12-09 10:44:55.611306871 +0000 UTC m=+3451.945811538" Dec 09 10:44:56 crc kubenswrapper[4824]: I1209 10:44:56.600263 4824 generic.go:334] "Generic (PLEG): container finished" podID="cc82131a-c612-48b6-be27-495cebd6296b" containerID="f77652e6456ac553187dff866b75f91dea702397ccf685cdc1cb4dc06c532017" exitCode=0 Dec 09 10:44:56 crc kubenswrapper[4824]: I1209 10:44:56.600363 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9jnd8" event={"ID":"cc82131a-c612-48b6-be27-495cebd6296b","Type":"ContainerDied","Data":"f77652e6456ac553187dff866b75f91dea702397ccf685cdc1cb4dc06c532017"} Dec 09 10:44:57 crc kubenswrapper[4824]: I1209 10:44:57.613629 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9jnd8" event={"ID":"cc82131a-c612-48b6-be27-495cebd6296b","Type":"ContainerStarted","Data":"d1a6c60691fd7a0108b7dc69637e74903e5a743a47d71724bb65abec12c40372"} Dec 09 10:44:57 crc kubenswrapper[4824]: I1209 10:44:57.642317 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9jnd8" podStartSLOduration=3.107235217 podStartE2EDuration="5.642294667s" podCreationTimestamp="2025-12-09 10:44:52 +0000 UTC" firstStartedPulling="2025-12-09 10:44:54.570148091 +0000 UTC m=+3450.904652778" lastFinishedPulling="2025-12-09 10:44:57.105207561 +0000 UTC m=+3453.439712228" observedRunningTime="2025-12-09 10:44:57.635735885 +0000 UTC m=+3453.970240552" watchObservedRunningTime="2025-12-09 10:44:57.642294667 +0000 UTC m=+3453.976799334" Dec 09 10:44:59 crc kubenswrapper[4824]: I1209 10:44:59.110152 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wk72l" Dec 09 10:44:59 crc kubenswrapper[4824]: I1209 10:44:59.110405 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wk72l" Dec 09 10:45:00 crc kubenswrapper[4824]: I1209 10:45:00.163388 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wk72l" podUID="be531b94-496d-45d7-9b5e-6532bd4f9890" containerName="registry-server" probeResult="failure" output=< Dec 09 10:45:00 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 10:45:00 crc kubenswrapper[4824]: > Dec 09 10:45:00 crc kubenswrapper[4824]: I1209 10:45:00.180942 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421285-kq5vl"] Dec 09 10:45:00 crc kubenswrapper[4824]: I1209 10:45:00.183282 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421285-kq5vl" Dec 09 10:45:00 crc kubenswrapper[4824]: I1209 10:45:00.185898 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 10:45:00 crc kubenswrapper[4824]: I1209 10:45:00.190909 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 10:45:00 crc kubenswrapper[4824]: I1209 10:45:00.197120 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421285-kq5vl"] Dec 09 10:45:00 crc kubenswrapper[4824]: I1209 10:45:00.334227 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4e327a9-3e2c-4d99-b54b-2170a944e839-config-volume\") pod \"collect-profiles-29421285-kq5vl\" (UID: \"b4e327a9-3e2c-4d99-b54b-2170a944e839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421285-kq5vl" Dec 09 10:45:00 crc kubenswrapper[4824]: I1209 10:45:00.334417 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4e327a9-3e2c-4d99-b54b-2170a944e839-secret-volume\") pod \"collect-profiles-29421285-kq5vl\" (UID: \"b4e327a9-3e2c-4d99-b54b-2170a944e839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421285-kq5vl" Dec 09 10:45:00 crc kubenswrapper[4824]: I1209 10:45:00.334757 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87q7w\" (UniqueName: \"kubernetes.io/projected/b4e327a9-3e2c-4d99-b54b-2170a944e839-kube-api-access-87q7w\") pod \"collect-profiles-29421285-kq5vl\" (UID: \"b4e327a9-3e2c-4d99-b54b-2170a944e839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421285-kq5vl" Dec 09 10:45:00 crc kubenswrapper[4824]: I1209 10:45:00.436864 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4e327a9-3e2c-4d99-b54b-2170a944e839-secret-volume\") pod \"collect-profiles-29421285-kq5vl\" (UID: \"b4e327a9-3e2c-4d99-b54b-2170a944e839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421285-kq5vl" Dec 09 10:45:00 crc kubenswrapper[4824]: I1209 10:45:00.437029 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87q7w\" (UniqueName: \"kubernetes.io/projected/b4e327a9-3e2c-4d99-b54b-2170a944e839-kube-api-access-87q7w\") pod \"collect-profiles-29421285-kq5vl\" (UID: \"b4e327a9-3e2c-4d99-b54b-2170a944e839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421285-kq5vl" Dec 09 10:45:00 crc kubenswrapper[4824]: I1209 10:45:00.437144 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4e327a9-3e2c-4d99-b54b-2170a944e839-config-volume\") pod \"collect-profiles-29421285-kq5vl\" (UID: \"b4e327a9-3e2c-4d99-b54b-2170a944e839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421285-kq5vl" Dec 09 10:45:00 crc kubenswrapper[4824]: I1209 10:45:00.438144 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4e327a9-3e2c-4d99-b54b-2170a944e839-config-volume\") pod \"collect-profiles-29421285-kq5vl\" (UID: \"b4e327a9-3e2c-4d99-b54b-2170a944e839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421285-kq5vl" Dec 09 10:45:00 crc kubenswrapper[4824]: I1209 10:45:00.445355 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4e327a9-3e2c-4d99-b54b-2170a944e839-secret-volume\") pod \"collect-profiles-29421285-kq5vl\" (UID: \"b4e327a9-3e2c-4d99-b54b-2170a944e839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421285-kq5vl" Dec 09 10:45:00 crc kubenswrapper[4824]: I1209 10:45:00.455944 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87q7w\" (UniqueName: \"kubernetes.io/projected/b4e327a9-3e2c-4d99-b54b-2170a944e839-kube-api-access-87q7w\") pod \"collect-profiles-29421285-kq5vl\" (UID: \"b4e327a9-3e2c-4d99-b54b-2170a944e839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421285-kq5vl" Dec 09 10:45:00 crc kubenswrapper[4824]: I1209 10:45:00.514104 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421285-kq5vl" Dec 09 10:45:01 crc kubenswrapper[4824]: I1209 10:45:01.016470 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421285-kq5vl"] Dec 09 10:45:01 crc kubenswrapper[4824]: I1209 10:45:01.677243 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421285-kq5vl" event={"ID":"b4e327a9-3e2c-4d99-b54b-2170a944e839","Type":"ContainerStarted","Data":"74dac484b2b7d8214aca9fbf913a0fdc72d50b22fbdd7ee64aec416cc968e910"} Dec 09 10:45:01 crc kubenswrapper[4824]: I1209 10:45:01.677667 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421285-kq5vl" event={"ID":"b4e327a9-3e2c-4d99-b54b-2170a944e839","Type":"ContainerStarted","Data":"d33be3c0dfde35513b97d1cd3172bac9a7bd242771c3a3a6ce068561e86308ca"} Dec 09 10:45:01 crc kubenswrapper[4824]: I1209 10:45:01.698413 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29421285-kq5vl" podStartSLOduration=1.69838708 podStartE2EDuration="1.69838708s" podCreationTimestamp="2025-12-09 10:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 10:45:01.695637255 +0000 UTC m=+3458.030141932" watchObservedRunningTime="2025-12-09 10:45:01.69838708 +0000 UTC m=+3458.032891747" Dec 09 10:45:02 crc kubenswrapper[4824]: I1209 10:45:02.689679 4824 generic.go:334] "Generic (PLEG): container finished" podID="b4e327a9-3e2c-4d99-b54b-2170a944e839" containerID="74dac484b2b7d8214aca9fbf913a0fdc72d50b22fbdd7ee64aec416cc968e910" exitCode=0 Dec 09 10:45:02 crc kubenswrapper[4824]: I1209 10:45:02.690085 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421285-kq5vl" event={"ID":"b4e327a9-3e2c-4d99-b54b-2170a944e839","Type":"ContainerDied","Data":"74dac484b2b7d8214aca9fbf913a0fdc72d50b22fbdd7ee64aec416cc968e910"} Dec 09 10:45:02 crc kubenswrapper[4824]: I1209 10:45:02.993724 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9jnd8" Dec 09 10:45:02 crc kubenswrapper[4824]: I1209 10:45:02.993763 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9jnd8" Dec 09 10:45:03 crc kubenswrapper[4824]: I1209 10:45:03.048128 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9jnd8" Dec 09 10:45:03 crc kubenswrapper[4824]: I1209 10:45:03.756174 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9jnd8" Dec 09 10:45:03 crc kubenswrapper[4824]: I1209 10:45:03.813159 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9jnd8"] Dec 09 10:45:04 crc kubenswrapper[4824]: I1209 10:45:04.503090 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421285-kq5vl" Dec 09 10:45:04 crc kubenswrapper[4824]: I1209 10:45:04.549151 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87q7w\" (UniqueName: \"kubernetes.io/projected/b4e327a9-3e2c-4d99-b54b-2170a944e839-kube-api-access-87q7w\") pod \"b4e327a9-3e2c-4d99-b54b-2170a944e839\" (UID: \"b4e327a9-3e2c-4d99-b54b-2170a944e839\") " Dec 09 10:45:04 crc kubenswrapper[4824]: I1209 10:45:04.549381 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4e327a9-3e2c-4d99-b54b-2170a944e839-config-volume\") pod \"b4e327a9-3e2c-4d99-b54b-2170a944e839\" (UID: \"b4e327a9-3e2c-4d99-b54b-2170a944e839\") " Dec 09 10:45:04 crc kubenswrapper[4824]: I1209 10:45:04.549417 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4e327a9-3e2c-4d99-b54b-2170a944e839-secret-volume\") pod \"b4e327a9-3e2c-4d99-b54b-2170a944e839\" (UID: \"b4e327a9-3e2c-4d99-b54b-2170a944e839\") " Dec 09 10:45:04 crc kubenswrapper[4824]: I1209 10:45:04.550319 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4e327a9-3e2c-4d99-b54b-2170a944e839-config-volume" (OuterVolumeSpecName: "config-volume") pod "b4e327a9-3e2c-4d99-b54b-2170a944e839" (UID: "b4e327a9-3e2c-4d99-b54b-2170a944e839"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 10:45:04 crc kubenswrapper[4824]: I1209 10:45:04.556000 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4e327a9-3e2c-4d99-b54b-2170a944e839-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b4e327a9-3e2c-4d99-b54b-2170a944e839" (UID: "b4e327a9-3e2c-4d99-b54b-2170a944e839"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 10:45:04 crc kubenswrapper[4824]: I1209 10:45:04.556141 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4e327a9-3e2c-4d99-b54b-2170a944e839-kube-api-access-87q7w" (OuterVolumeSpecName: "kube-api-access-87q7w") pod "b4e327a9-3e2c-4d99-b54b-2170a944e839" (UID: "b4e327a9-3e2c-4d99-b54b-2170a944e839"). InnerVolumeSpecName "kube-api-access-87q7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:45:04 crc kubenswrapper[4824]: I1209 10:45:04.652721 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87q7w\" (UniqueName: \"kubernetes.io/projected/b4e327a9-3e2c-4d99-b54b-2170a944e839-kube-api-access-87q7w\") on node \"crc\" DevicePath \"\"" Dec 09 10:45:04 crc kubenswrapper[4824]: I1209 10:45:04.652758 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4e327a9-3e2c-4d99-b54b-2170a944e839-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 10:45:04 crc kubenswrapper[4824]: I1209 10:45:04.652930 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4e327a9-3e2c-4d99-b54b-2170a944e839-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 10:45:04 crc kubenswrapper[4824]: I1209 10:45:04.713174 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421285-kq5vl" Dec 09 10:45:04 crc kubenswrapper[4824]: I1209 10:45:04.713132 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421285-kq5vl" event={"ID":"b4e327a9-3e2c-4d99-b54b-2170a944e839","Type":"ContainerDied","Data":"d33be3c0dfde35513b97d1cd3172bac9a7bd242771c3a3a6ce068561e86308ca"} Dec 09 10:45:04 crc kubenswrapper[4824]: I1209 10:45:04.713562 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d33be3c0dfde35513b97d1cd3172bac9a7bd242771c3a3a6ce068561e86308ca" Dec 09 10:45:04 crc kubenswrapper[4824]: I1209 10:45:04.779792 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421240-cqlx2"] Dec 09 10:45:04 crc kubenswrapper[4824]: I1209 10:45:04.795385 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421240-cqlx2"] Dec 09 10:45:05 crc kubenswrapper[4824]: I1209 10:45:05.726858 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9jnd8" podUID="cc82131a-c612-48b6-be27-495cebd6296b" containerName="registry-server" containerID="cri-o://d1a6c60691fd7a0108b7dc69637e74903e5a743a47d71724bb65abec12c40372" gracePeriod=2 Dec 09 10:45:05 crc kubenswrapper[4824]: I1209 10:45:05.929862 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ca530df-dd8f-427e-bc32-6b107395b6d5" path="/var/lib/kubelet/pods/4ca530df-dd8f-427e-bc32-6b107395b6d5/volumes" Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.271545 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9jnd8" Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.299983 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc82131a-c612-48b6-be27-495cebd6296b-utilities\") pod \"cc82131a-c612-48b6-be27-495cebd6296b\" (UID: \"cc82131a-c612-48b6-be27-495cebd6296b\") " Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.300153 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc82131a-c612-48b6-be27-495cebd6296b-catalog-content\") pod \"cc82131a-c612-48b6-be27-495cebd6296b\" (UID: \"cc82131a-c612-48b6-be27-495cebd6296b\") " Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.300251 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klhl9\" (UniqueName: \"kubernetes.io/projected/cc82131a-c612-48b6-be27-495cebd6296b-kube-api-access-klhl9\") pod \"cc82131a-c612-48b6-be27-495cebd6296b\" (UID: \"cc82131a-c612-48b6-be27-495cebd6296b\") " Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.301056 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc82131a-c612-48b6-be27-495cebd6296b-utilities" (OuterVolumeSpecName: "utilities") pod "cc82131a-c612-48b6-be27-495cebd6296b" (UID: "cc82131a-c612-48b6-be27-495cebd6296b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.311382 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc82131a-c612-48b6-be27-495cebd6296b-kube-api-access-klhl9" (OuterVolumeSpecName: "kube-api-access-klhl9") pod "cc82131a-c612-48b6-be27-495cebd6296b" (UID: "cc82131a-c612-48b6-be27-495cebd6296b"). InnerVolumeSpecName "kube-api-access-klhl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.349384 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc82131a-c612-48b6-be27-495cebd6296b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc82131a-c612-48b6-be27-495cebd6296b" (UID: "cc82131a-c612-48b6-be27-495cebd6296b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.403735 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc82131a-c612-48b6-be27-495cebd6296b-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.404086 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc82131a-c612-48b6-be27-495cebd6296b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.404189 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klhl9\" (UniqueName: \"kubernetes.io/projected/cc82131a-c612-48b6-be27-495cebd6296b-kube-api-access-klhl9\") on node \"crc\" DevicePath \"\"" Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.741259 4824 generic.go:334] "Generic (PLEG): container finished" podID="cc82131a-c612-48b6-be27-495cebd6296b" containerID="d1a6c60691fd7a0108b7dc69637e74903e5a743a47d71724bb65abec12c40372" exitCode=0 Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.741317 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9jnd8" event={"ID":"cc82131a-c612-48b6-be27-495cebd6296b","Type":"ContainerDied","Data":"d1a6c60691fd7a0108b7dc69637e74903e5a743a47d71724bb65abec12c40372"} Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.741660 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9jnd8" event={"ID":"cc82131a-c612-48b6-be27-495cebd6296b","Type":"ContainerDied","Data":"f330bf70af3fd4b6c197529c1648f74c8e5d63743d843e9b0bc5c1d75e4ec060"} Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.741390 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9jnd8" Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.741715 4824 scope.go:117] "RemoveContainer" containerID="d1a6c60691fd7a0108b7dc69637e74903e5a743a47d71724bb65abec12c40372" Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.777930 4824 scope.go:117] "RemoveContainer" containerID="f77652e6456ac553187dff866b75f91dea702397ccf685cdc1cb4dc06c532017" Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.788573 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9jnd8"] Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.800837 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9jnd8"] Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.813338 4824 scope.go:117] "RemoveContainer" containerID="3620580e79bb5e72eef0a1cec5838c5aed128251bcf9ca3c97fbeebbbd4d8420" Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.865861 4824 scope.go:117] "RemoveContainer" containerID="d1a6c60691fd7a0108b7dc69637e74903e5a743a47d71724bb65abec12c40372" Dec 09 10:45:06 crc kubenswrapper[4824]: E1209 10:45:06.866579 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1a6c60691fd7a0108b7dc69637e74903e5a743a47d71724bb65abec12c40372\": container with ID starting with d1a6c60691fd7a0108b7dc69637e74903e5a743a47d71724bb65abec12c40372 not found: ID does not exist" containerID="d1a6c60691fd7a0108b7dc69637e74903e5a743a47d71724bb65abec12c40372" Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.866632 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1a6c60691fd7a0108b7dc69637e74903e5a743a47d71724bb65abec12c40372"} err="failed to get container status \"d1a6c60691fd7a0108b7dc69637e74903e5a743a47d71724bb65abec12c40372\": rpc error: code = NotFound desc = could not find container \"d1a6c60691fd7a0108b7dc69637e74903e5a743a47d71724bb65abec12c40372\": container with ID starting with d1a6c60691fd7a0108b7dc69637e74903e5a743a47d71724bb65abec12c40372 not found: ID does not exist" Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.866665 4824 scope.go:117] "RemoveContainer" containerID="f77652e6456ac553187dff866b75f91dea702397ccf685cdc1cb4dc06c532017" Dec 09 10:45:06 crc kubenswrapper[4824]: E1209 10:45:06.867273 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f77652e6456ac553187dff866b75f91dea702397ccf685cdc1cb4dc06c532017\": container with ID starting with f77652e6456ac553187dff866b75f91dea702397ccf685cdc1cb4dc06c532017 not found: ID does not exist" containerID="f77652e6456ac553187dff866b75f91dea702397ccf685cdc1cb4dc06c532017" Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.867309 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f77652e6456ac553187dff866b75f91dea702397ccf685cdc1cb4dc06c532017"} err="failed to get container status \"f77652e6456ac553187dff866b75f91dea702397ccf685cdc1cb4dc06c532017\": rpc error: code = NotFound desc = could not find container \"f77652e6456ac553187dff866b75f91dea702397ccf685cdc1cb4dc06c532017\": container with ID starting with f77652e6456ac553187dff866b75f91dea702397ccf685cdc1cb4dc06c532017 not found: ID does not exist" Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.867332 4824 scope.go:117] "RemoveContainer" containerID="3620580e79bb5e72eef0a1cec5838c5aed128251bcf9ca3c97fbeebbbd4d8420" Dec 09 10:45:06 crc kubenswrapper[4824]: E1209 10:45:06.867626 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3620580e79bb5e72eef0a1cec5838c5aed128251bcf9ca3c97fbeebbbd4d8420\": container with ID starting with 3620580e79bb5e72eef0a1cec5838c5aed128251bcf9ca3c97fbeebbbd4d8420 not found: ID does not exist" containerID="3620580e79bb5e72eef0a1cec5838c5aed128251bcf9ca3c97fbeebbbd4d8420" Dec 09 10:45:06 crc kubenswrapper[4824]: I1209 10:45:06.867655 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3620580e79bb5e72eef0a1cec5838c5aed128251bcf9ca3c97fbeebbbd4d8420"} err="failed to get container status \"3620580e79bb5e72eef0a1cec5838c5aed128251bcf9ca3c97fbeebbbd4d8420\": rpc error: code = NotFound desc = could not find container \"3620580e79bb5e72eef0a1cec5838c5aed128251bcf9ca3c97fbeebbbd4d8420\": container with ID starting with 3620580e79bb5e72eef0a1cec5838c5aed128251bcf9ca3c97fbeebbbd4d8420 not found: ID does not exist" Dec 09 10:45:07 crc kubenswrapper[4824]: I1209 10:45:07.924854 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc82131a-c612-48b6-be27-495cebd6296b" path="/var/lib/kubelet/pods/cc82131a-c612-48b6-be27-495cebd6296b/volumes" Dec 09 10:45:09 crc kubenswrapper[4824]: I1209 10:45:09.161618 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wk72l" Dec 09 10:45:09 crc kubenswrapper[4824]: I1209 10:45:09.213130 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wk72l" Dec 09 10:45:10 crc kubenswrapper[4824]: I1209 10:45:10.889484 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wk72l"] Dec 09 10:45:10 crc kubenswrapper[4824]: I1209 10:45:10.890021 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wk72l" podUID="be531b94-496d-45d7-9b5e-6532bd4f9890" containerName="registry-server" containerID="cri-o://271b56f2e4ef0073d0c4f22851310329a0bbc483f52691cd44c45164e60467c2" gracePeriod=2 Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.423926 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wk72l" Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.430959 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be531b94-496d-45d7-9b5e-6532bd4f9890-catalog-content\") pod \"be531b94-496d-45d7-9b5e-6532bd4f9890\" (UID: \"be531b94-496d-45d7-9b5e-6532bd4f9890\") " Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.431051 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qc8b\" (UniqueName: \"kubernetes.io/projected/be531b94-496d-45d7-9b5e-6532bd4f9890-kube-api-access-9qc8b\") pod \"be531b94-496d-45d7-9b5e-6532bd4f9890\" (UID: \"be531b94-496d-45d7-9b5e-6532bd4f9890\") " Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.431111 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be531b94-496d-45d7-9b5e-6532bd4f9890-utilities\") pod \"be531b94-496d-45d7-9b5e-6532bd4f9890\" (UID: \"be531b94-496d-45d7-9b5e-6532bd4f9890\") " Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.432151 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be531b94-496d-45d7-9b5e-6532bd4f9890-utilities" (OuterVolumeSpecName: "utilities") pod "be531b94-496d-45d7-9b5e-6532bd4f9890" (UID: "be531b94-496d-45d7-9b5e-6532bd4f9890"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.436812 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be531b94-496d-45d7-9b5e-6532bd4f9890-kube-api-access-9qc8b" (OuterVolumeSpecName: "kube-api-access-9qc8b") pod "be531b94-496d-45d7-9b5e-6532bd4f9890" (UID: "be531b94-496d-45d7-9b5e-6532bd4f9890"). InnerVolumeSpecName "kube-api-access-9qc8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.533987 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qc8b\" (UniqueName: \"kubernetes.io/projected/be531b94-496d-45d7-9b5e-6532bd4f9890-kube-api-access-9qc8b\") on node \"crc\" DevicePath \"\"" Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.534211 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be531b94-496d-45d7-9b5e-6532bd4f9890-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.554366 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be531b94-496d-45d7-9b5e-6532bd4f9890-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be531b94-496d-45d7-9b5e-6532bd4f9890" (UID: "be531b94-496d-45d7-9b5e-6532bd4f9890"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.636871 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be531b94-496d-45d7-9b5e-6532bd4f9890-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.801817 4824 generic.go:334] "Generic (PLEG): container finished" podID="be531b94-496d-45d7-9b5e-6532bd4f9890" containerID="271b56f2e4ef0073d0c4f22851310329a0bbc483f52691cd44c45164e60467c2" exitCode=0 Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.801865 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wk72l" event={"ID":"be531b94-496d-45d7-9b5e-6532bd4f9890","Type":"ContainerDied","Data":"271b56f2e4ef0073d0c4f22851310329a0bbc483f52691cd44c45164e60467c2"} Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.801892 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wk72l" event={"ID":"be531b94-496d-45d7-9b5e-6532bd4f9890","Type":"ContainerDied","Data":"6b2717a3ec5ed885482dcee97baf865eae8c26f632a8159c532c9d4cd33c0e7a"} Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.801888 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wk72l" Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.801909 4824 scope.go:117] "RemoveContainer" containerID="271b56f2e4ef0073d0c4f22851310329a0bbc483f52691cd44c45164e60467c2" Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.831821 4824 scope.go:117] "RemoveContainer" containerID="995fa5682ba3e4e72b03173615f379ded0568046a23780ec66d954cc43143534" Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.843659 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wk72l"] Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.859457 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wk72l"] Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.872496 4824 scope.go:117] "RemoveContainer" containerID="9f6b541c1d3e8df40a30f6c785c1fba074b6f3db4808b71565d62c29f578229e" Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.920895 4824 scope.go:117] "RemoveContainer" containerID="271b56f2e4ef0073d0c4f22851310329a0bbc483f52691cd44c45164e60467c2" Dec 09 10:45:11 crc kubenswrapper[4824]: E1209 10:45:11.921418 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"271b56f2e4ef0073d0c4f22851310329a0bbc483f52691cd44c45164e60467c2\": container with ID starting with 271b56f2e4ef0073d0c4f22851310329a0bbc483f52691cd44c45164e60467c2 not found: ID does not exist" containerID="271b56f2e4ef0073d0c4f22851310329a0bbc483f52691cd44c45164e60467c2" Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.921461 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"271b56f2e4ef0073d0c4f22851310329a0bbc483f52691cd44c45164e60467c2"} err="failed to get container status \"271b56f2e4ef0073d0c4f22851310329a0bbc483f52691cd44c45164e60467c2\": rpc error: code = NotFound desc = could not find container \"271b56f2e4ef0073d0c4f22851310329a0bbc483f52691cd44c45164e60467c2\": container with ID starting with 271b56f2e4ef0073d0c4f22851310329a0bbc483f52691cd44c45164e60467c2 not found: ID does not exist" Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.921486 4824 scope.go:117] "RemoveContainer" containerID="995fa5682ba3e4e72b03173615f379ded0568046a23780ec66d954cc43143534" Dec 09 10:45:11 crc kubenswrapper[4824]: E1209 10:45:11.921767 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"995fa5682ba3e4e72b03173615f379ded0568046a23780ec66d954cc43143534\": container with ID starting with 995fa5682ba3e4e72b03173615f379ded0568046a23780ec66d954cc43143534 not found: ID does not exist" containerID="995fa5682ba3e4e72b03173615f379ded0568046a23780ec66d954cc43143534" Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.921824 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"995fa5682ba3e4e72b03173615f379ded0568046a23780ec66d954cc43143534"} err="failed to get container status \"995fa5682ba3e4e72b03173615f379ded0568046a23780ec66d954cc43143534\": rpc error: code = NotFound desc = could not find container \"995fa5682ba3e4e72b03173615f379ded0568046a23780ec66d954cc43143534\": container with ID starting with 995fa5682ba3e4e72b03173615f379ded0568046a23780ec66d954cc43143534 not found: ID does not exist" Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.921842 4824 scope.go:117] "RemoveContainer" containerID="9f6b541c1d3e8df40a30f6c785c1fba074b6f3db4808b71565d62c29f578229e" Dec 09 10:45:11 crc kubenswrapper[4824]: E1209 10:45:11.922074 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f6b541c1d3e8df40a30f6c785c1fba074b6f3db4808b71565d62c29f578229e\": container with ID starting with 9f6b541c1d3e8df40a30f6c785c1fba074b6f3db4808b71565d62c29f578229e not found: ID does not exist" containerID="9f6b541c1d3e8df40a30f6c785c1fba074b6f3db4808b71565d62c29f578229e" Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.922103 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f6b541c1d3e8df40a30f6c785c1fba074b6f3db4808b71565d62c29f578229e"} err="failed to get container status \"9f6b541c1d3e8df40a30f6c785c1fba074b6f3db4808b71565d62c29f578229e\": rpc error: code = NotFound desc = could not find container \"9f6b541c1d3e8df40a30f6c785c1fba074b6f3db4808b71565d62c29f578229e\": container with ID starting with 9f6b541c1d3e8df40a30f6c785c1fba074b6f3db4808b71565d62c29f578229e not found: ID does not exist" Dec 09 10:45:11 crc kubenswrapper[4824]: I1209 10:45:11.925105 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be531b94-496d-45d7-9b5e-6532bd4f9890" path="/var/lib/kubelet/pods/be531b94-496d-45d7-9b5e-6532bd4f9890/volumes" Dec 09 10:45:18 crc kubenswrapper[4824]: I1209 10:45:18.491446 4824 scope.go:117] "RemoveContainer" containerID="448b6595c82b3a5d4f873fbd2d468b683458cbd4e6c76699fac8e8505395051b" Dec 09 10:47:02 crc kubenswrapper[4824]: I1209 10:47:02.861397 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:47:02 crc kubenswrapper[4824]: I1209 10:47:02.862017 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:47:32 crc kubenswrapper[4824]: I1209 10:47:32.861569 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:47:32 crc kubenswrapper[4824]: I1209 10:47:32.863193 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:48:02 crc kubenswrapper[4824]: I1209 10:48:02.861666 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:48:02 crc kubenswrapper[4824]: I1209 10:48:02.863174 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:48:02 crc kubenswrapper[4824]: I1209 10:48:02.863277 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 10:48:02 crc kubenswrapper[4824]: I1209 10:48:02.864394 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7050a696e11544a6263f8787d7539e141eb5adea02be70809cc2757af296657c"} pod="openshift-machine-config-operator/machine-config-daemon-dth8x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 10:48:02 crc kubenswrapper[4824]: I1209 10:48:02.864461 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" containerID="cri-o://7050a696e11544a6263f8787d7539e141eb5adea02be70809cc2757af296657c" gracePeriod=600 Dec 09 10:48:03 crc kubenswrapper[4824]: I1209 10:48:03.124157 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerDied","Data":"7050a696e11544a6263f8787d7539e141eb5adea02be70809cc2757af296657c"} Dec 09 10:48:03 crc kubenswrapper[4824]: I1209 10:48:03.124490 4824 scope.go:117] "RemoveContainer" containerID="ae742a965243fabcb31e789e17faf2f81df786a0063e7ef19d3743c627548fcb" Dec 09 10:48:03 crc kubenswrapper[4824]: I1209 10:48:03.124111 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerID="7050a696e11544a6263f8787d7539e141eb5adea02be70809cc2757af296657c" exitCode=0 Dec 09 10:48:04 crc kubenswrapper[4824]: I1209 10:48:04.162255 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerStarted","Data":"fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e"} Dec 09 10:50:32 crc kubenswrapper[4824]: I1209 10:50:32.861064 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:50:32 crc kubenswrapper[4824]: I1209 10:50:32.861670 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:50:34 crc kubenswrapper[4824]: I1209 10:50:34.787322 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qzdq8"] Dec 09 10:50:34 crc kubenswrapper[4824]: E1209 10:50:34.790048 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc82131a-c612-48b6-be27-495cebd6296b" containerName="extract-content" Dec 09 10:50:34 crc kubenswrapper[4824]: I1209 10:50:34.790193 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc82131a-c612-48b6-be27-495cebd6296b" containerName="extract-content" Dec 09 10:50:34 crc kubenswrapper[4824]: E1209 10:50:34.790316 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc82131a-c612-48b6-be27-495cebd6296b" containerName="extract-utilities" Dec 09 10:50:34 crc kubenswrapper[4824]: I1209 10:50:34.790411 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc82131a-c612-48b6-be27-495cebd6296b" containerName="extract-utilities" Dec 09 10:50:34 crc kubenswrapper[4824]: E1209 10:50:34.790586 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be531b94-496d-45d7-9b5e-6532bd4f9890" containerName="extract-content" Dec 09 10:50:34 crc kubenswrapper[4824]: I1209 10:50:34.790726 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="be531b94-496d-45d7-9b5e-6532bd4f9890" containerName="extract-content" Dec 09 10:50:34 crc kubenswrapper[4824]: E1209 10:50:34.790896 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be531b94-496d-45d7-9b5e-6532bd4f9890" containerName="extract-utilities" Dec 09 10:50:34 crc kubenswrapper[4824]: I1209 10:50:34.791017 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="be531b94-496d-45d7-9b5e-6532bd4f9890" containerName="extract-utilities" Dec 09 10:50:34 crc kubenswrapper[4824]: E1209 10:50:34.791137 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be531b94-496d-45d7-9b5e-6532bd4f9890" containerName="registry-server" Dec 09 10:50:34 crc kubenswrapper[4824]: I1209 10:50:34.791255 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="be531b94-496d-45d7-9b5e-6532bd4f9890" containerName="registry-server" Dec 09 10:50:34 crc kubenswrapper[4824]: E1209 10:50:34.791392 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4e327a9-3e2c-4d99-b54b-2170a944e839" containerName="collect-profiles" Dec 09 10:50:34 crc kubenswrapper[4824]: I1209 10:50:34.791518 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4e327a9-3e2c-4d99-b54b-2170a944e839" containerName="collect-profiles" Dec 09 10:50:34 crc kubenswrapper[4824]: E1209 10:50:34.791637 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc82131a-c612-48b6-be27-495cebd6296b" containerName="registry-server" Dec 09 10:50:34 crc kubenswrapper[4824]: I1209 10:50:34.791727 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc82131a-c612-48b6-be27-495cebd6296b" containerName="registry-server" Dec 09 10:50:34 crc kubenswrapper[4824]: I1209 10:50:34.792230 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc82131a-c612-48b6-be27-495cebd6296b" containerName="registry-server" Dec 09 10:50:34 crc kubenswrapper[4824]: I1209 10:50:34.792332 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4e327a9-3e2c-4d99-b54b-2170a944e839" containerName="collect-profiles" Dec 09 10:50:34 crc kubenswrapper[4824]: I1209 10:50:34.792516 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="be531b94-496d-45d7-9b5e-6532bd4f9890" containerName="registry-server" Dec 09 10:50:34 crc kubenswrapper[4824]: I1209 10:50:34.795067 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qzdq8" Dec 09 10:50:34 crc kubenswrapper[4824]: I1209 10:50:34.799571 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzdq8"] Dec 09 10:50:34 crc kubenswrapper[4824]: I1209 10:50:34.858609 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ab7caf8-ce36-4db9-bc21-7d33520b96d1-utilities\") pod \"redhat-marketplace-qzdq8\" (UID: \"2ab7caf8-ce36-4db9-bc21-7d33520b96d1\") " pod="openshift-marketplace/redhat-marketplace-qzdq8" Dec 09 10:50:34 crc kubenswrapper[4824]: I1209 10:50:34.858660 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klvm8\" (UniqueName: \"kubernetes.io/projected/2ab7caf8-ce36-4db9-bc21-7d33520b96d1-kube-api-access-klvm8\") pod \"redhat-marketplace-qzdq8\" (UID: \"2ab7caf8-ce36-4db9-bc21-7d33520b96d1\") " pod="openshift-marketplace/redhat-marketplace-qzdq8" Dec 09 10:50:34 crc kubenswrapper[4824]: I1209 10:50:34.858747 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ab7caf8-ce36-4db9-bc21-7d33520b96d1-catalog-content\") pod \"redhat-marketplace-qzdq8\" (UID: \"2ab7caf8-ce36-4db9-bc21-7d33520b96d1\") " pod="openshift-marketplace/redhat-marketplace-qzdq8" Dec 09 10:50:34 crc kubenswrapper[4824]: I1209 10:50:34.960907 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ab7caf8-ce36-4db9-bc21-7d33520b96d1-utilities\") pod \"redhat-marketplace-qzdq8\" (UID: \"2ab7caf8-ce36-4db9-bc21-7d33520b96d1\") " pod="openshift-marketplace/redhat-marketplace-qzdq8" Dec 09 10:50:34 crc kubenswrapper[4824]: I1209 10:50:34.960959 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klvm8\" (UniqueName: \"kubernetes.io/projected/2ab7caf8-ce36-4db9-bc21-7d33520b96d1-kube-api-access-klvm8\") pod \"redhat-marketplace-qzdq8\" (UID: \"2ab7caf8-ce36-4db9-bc21-7d33520b96d1\") " pod="openshift-marketplace/redhat-marketplace-qzdq8" Dec 09 10:50:34 crc kubenswrapper[4824]: I1209 10:50:34.961059 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ab7caf8-ce36-4db9-bc21-7d33520b96d1-catalog-content\") pod \"redhat-marketplace-qzdq8\" (UID: \"2ab7caf8-ce36-4db9-bc21-7d33520b96d1\") " pod="openshift-marketplace/redhat-marketplace-qzdq8" Dec 09 10:50:34 crc kubenswrapper[4824]: I1209 10:50:34.961607 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ab7caf8-ce36-4db9-bc21-7d33520b96d1-utilities\") pod \"redhat-marketplace-qzdq8\" (UID: \"2ab7caf8-ce36-4db9-bc21-7d33520b96d1\") " pod="openshift-marketplace/redhat-marketplace-qzdq8" Dec 09 10:50:34 crc kubenswrapper[4824]: I1209 10:50:34.961633 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ab7caf8-ce36-4db9-bc21-7d33520b96d1-catalog-content\") pod \"redhat-marketplace-qzdq8\" (UID: \"2ab7caf8-ce36-4db9-bc21-7d33520b96d1\") " pod="openshift-marketplace/redhat-marketplace-qzdq8" Dec 09 10:50:34 crc kubenswrapper[4824]: I1209 10:50:34.983544 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klvm8\" (UniqueName: \"kubernetes.io/projected/2ab7caf8-ce36-4db9-bc21-7d33520b96d1-kube-api-access-klvm8\") pod \"redhat-marketplace-qzdq8\" (UID: \"2ab7caf8-ce36-4db9-bc21-7d33520b96d1\") " pod="openshift-marketplace/redhat-marketplace-qzdq8" Dec 09 10:50:35 crc kubenswrapper[4824]: I1209 10:50:35.156743 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qzdq8" Dec 09 10:50:35 crc kubenswrapper[4824]: I1209 10:50:35.687535 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzdq8"] Dec 09 10:50:36 crc kubenswrapper[4824]: I1209 10:50:36.126586 4824 generic.go:334] "Generic (PLEG): container finished" podID="2ab7caf8-ce36-4db9-bc21-7d33520b96d1" containerID="69a07c36ebf55bfc33838e5dbf680813a0350b8327a88cb9583e1faacb9ba200" exitCode=0 Dec 09 10:50:36 crc kubenswrapper[4824]: I1209 10:50:36.126665 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzdq8" event={"ID":"2ab7caf8-ce36-4db9-bc21-7d33520b96d1","Type":"ContainerDied","Data":"69a07c36ebf55bfc33838e5dbf680813a0350b8327a88cb9583e1faacb9ba200"} Dec 09 10:50:36 crc kubenswrapper[4824]: I1209 10:50:36.126742 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzdq8" event={"ID":"2ab7caf8-ce36-4db9-bc21-7d33520b96d1","Type":"ContainerStarted","Data":"ecf49845b8335147afa92b37a59fa761ed7c9cc972f508356713707e21e14301"} Dec 09 10:50:36 crc kubenswrapper[4824]: I1209 10:50:36.129742 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 10:50:37 crc kubenswrapper[4824]: I1209 10:50:37.137996 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzdq8" event={"ID":"2ab7caf8-ce36-4db9-bc21-7d33520b96d1","Type":"ContainerStarted","Data":"95a5f64697925ae7ef4bf6a2eeabee3f48855304f7f56c47a59fa46d369e1a69"} Dec 09 10:50:38 crc kubenswrapper[4824]: I1209 10:50:38.151673 4824 generic.go:334] "Generic (PLEG): container finished" podID="2ab7caf8-ce36-4db9-bc21-7d33520b96d1" containerID="95a5f64697925ae7ef4bf6a2eeabee3f48855304f7f56c47a59fa46d369e1a69" exitCode=0 Dec 09 10:50:38 crc kubenswrapper[4824]: I1209 10:50:38.151716 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzdq8" event={"ID":"2ab7caf8-ce36-4db9-bc21-7d33520b96d1","Type":"ContainerDied","Data":"95a5f64697925ae7ef4bf6a2eeabee3f48855304f7f56c47a59fa46d369e1a69"} Dec 09 10:50:39 crc kubenswrapper[4824]: I1209 10:50:39.162750 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzdq8" event={"ID":"2ab7caf8-ce36-4db9-bc21-7d33520b96d1","Type":"ContainerStarted","Data":"04e09f0957bbe091d0a4bdb9e6ea0ce46e8001ce1b20b6c48a4612c0bdfbf4ca"} Dec 09 10:50:39 crc kubenswrapper[4824]: I1209 10:50:39.194318 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qzdq8" podStartSLOduration=2.6422561890000003 podStartE2EDuration="5.194266874s" podCreationTimestamp="2025-12-09 10:50:34 +0000 UTC" firstStartedPulling="2025-12-09 10:50:36.12928587 +0000 UTC m=+3792.463790537" lastFinishedPulling="2025-12-09 10:50:38.681296555 +0000 UTC m=+3795.015801222" observedRunningTime="2025-12-09 10:50:39.182406906 +0000 UTC m=+3795.516911573" watchObservedRunningTime="2025-12-09 10:50:39.194266874 +0000 UTC m=+3795.528771541" Dec 09 10:50:45 crc kubenswrapper[4824]: I1209 10:50:45.157214 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qzdq8" Dec 09 10:50:45 crc kubenswrapper[4824]: I1209 10:50:45.157794 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qzdq8" Dec 09 10:50:45 crc kubenswrapper[4824]: I1209 10:50:45.205568 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qzdq8" Dec 09 10:50:45 crc kubenswrapper[4824]: I1209 10:50:45.296891 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qzdq8" Dec 09 10:50:45 crc kubenswrapper[4824]: I1209 10:50:45.464752 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzdq8"] Dec 09 10:50:47 crc kubenswrapper[4824]: I1209 10:50:47.264106 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qzdq8" podUID="2ab7caf8-ce36-4db9-bc21-7d33520b96d1" containerName="registry-server" containerID="cri-o://04e09f0957bbe091d0a4bdb9e6ea0ce46e8001ce1b20b6c48a4612c0bdfbf4ca" gracePeriod=2 Dec 09 10:50:47 crc kubenswrapper[4824]: I1209 10:50:47.784841 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qzdq8" Dec 09 10:50:47 crc kubenswrapper[4824]: I1209 10:50:47.792860 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ab7caf8-ce36-4db9-bc21-7d33520b96d1-utilities\") pod \"2ab7caf8-ce36-4db9-bc21-7d33520b96d1\" (UID: \"2ab7caf8-ce36-4db9-bc21-7d33520b96d1\") " Dec 09 10:50:47 crc kubenswrapper[4824]: I1209 10:50:47.792967 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klvm8\" (UniqueName: \"kubernetes.io/projected/2ab7caf8-ce36-4db9-bc21-7d33520b96d1-kube-api-access-klvm8\") pod \"2ab7caf8-ce36-4db9-bc21-7d33520b96d1\" (UID: \"2ab7caf8-ce36-4db9-bc21-7d33520b96d1\") " Dec 09 10:50:47 crc kubenswrapper[4824]: I1209 10:50:47.793102 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ab7caf8-ce36-4db9-bc21-7d33520b96d1-catalog-content\") pod \"2ab7caf8-ce36-4db9-bc21-7d33520b96d1\" (UID: \"2ab7caf8-ce36-4db9-bc21-7d33520b96d1\") " Dec 09 10:50:47 crc kubenswrapper[4824]: I1209 10:50:47.798574 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ab7caf8-ce36-4db9-bc21-7d33520b96d1-utilities" (OuterVolumeSpecName: "utilities") pod "2ab7caf8-ce36-4db9-bc21-7d33520b96d1" (UID: "2ab7caf8-ce36-4db9-bc21-7d33520b96d1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:50:47 crc kubenswrapper[4824]: I1209 10:50:47.805883 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ab7caf8-ce36-4db9-bc21-7d33520b96d1-kube-api-access-klvm8" (OuterVolumeSpecName: "kube-api-access-klvm8") pod "2ab7caf8-ce36-4db9-bc21-7d33520b96d1" (UID: "2ab7caf8-ce36-4db9-bc21-7d33520b96d1"). InnerVolumeSpecName "kube-api-access-klvm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:50:47 crc kubenswrapper[4824]: I1209 10:50:47.817537 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ab7caf8-ce36-4db9-bc21-7d33520b96d1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2ab7caf8-ce36-4db9-bc21-7d33520b96d1" (UID: "2ab7caf8-ce36-4db9-bc21-7d33520b96d1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:50:47 crc kubenswrapper[4824]: I1209 10:50:47.895879 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ab7caf8-ce36-4db9-bc21-7d33520b96d1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 10:50:47 crc kubenswrapper[4824]: I1209 10:50:47.895914 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ab7caf8-ce36-4db9-bc21-7d33520b96d1-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 10:50:47 crc kubenswrapper[4824]: I1209 10:50:47.895928 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klvm8\" (UniqueName: \"kubernetes.io/projected/2ab7caf8-ce36-4db9-bc21-7d33520b96d1-kube-api-access-klvm8\") on node \"crc\" DevicePath \"\"" Dec 09 10:50:48 crc kubenswrapper[4824]: I1209 10:50:48.277211 4824 generic.go:334] "Generic (PLEG): container finished" podID="2ab7caf8-ce36-4db9-bc21-7d33520b96d1" containerID="04e09f0957bbe091d0a4bdb9e6ea0ce46e8001ce1b20b6c48a4612c0bdfbf4ca" exitCode=0 Dec 09 10:50:48 crc kubenswrapper[4824]: I1209 10:50:48.277271 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qzdq8" Dec 09 10:50:48 crc kubenswrapper[4824]: I1209 10:50:48.277265 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzdq8" event={"ID":"2ab7caf8-ce36-4db9-bc21-7d33520b96d1","Type":"ContainerDied","Data":"04e09f0957bbe091d0a4bdb9e6ea0ce46e8001ce1b20b6c48a4612c0bdfbf4ca"} Dec 09 10:50:48 crc kubenswrapper[4824]: I1209 10:50:48.277414 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzdq8" event={"ID":"2ab7caf8-ce36-4db9-bc21-7d33520b96d1","Type":"ContainerDied","Data":"ecf49845b8335147afa92b37a59fa761ed7c9cc972f508356713707e21e14301"} Dec 09 10:50:48 crc kubenswrapper[4824]: I1209 10:50:48.277454 4824 scope.go:117] "RemoveContainer" containerID="04e09f0957bbe091d0a4bdb9e6ea0ce46e8001ce1b20b6c48a4612c0bdfbf4ca" Dec 09 10:50:48 crc kubenswrapper[4824]: I1209 10:50:48.305762 4824 scope.go:117] "RemoveContainer" containerID="95a5f64697925ae7ef4bf6a2eeabee3f48855304f7f56c47a59fa46d369e1a69" Dec 09 10:50:48 crc kubenswrapper[4824]: I1209 10:50:48.309085 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzdq8"] Dec 09 10:50:48 crc kubenswrapper[4824]: I1209 10:50:48.322789 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzdq8"] Dec 09 10:50:48 crc kubenswrapper[4824]: I1209 10:50:48.327344 4824 scope.go:117] "RemoveContainer" containerID="69a07c36ebf55bfc33838e5dbf680813a0350b8327a88cb9583e1faacb9ba200" Dec 09 10:50:48 crc kubenswrapper[4824]: I1209 10:50:48.387139 4824 scope.go:117] "RemoveContainer" containerID="04e09f0957bbe091d0a4bdb9e6ea0ce46e8001ce1b20b6c48a4612c0bdfbf4ca" Dec 09 10:50:48 crc kubenswrapper[4824]: E1209 10:50:48.387566 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04e09f0957bbe091d0a4bdb9e6ea0ce46e8001ce1b20b6c48a4612c0bdfbf4ca\": container with ID starting with 04e09f0957bbe091d0a4bdb9e6ea0ce46e8001ce1b20b6c48a4612c0bdfbf4ca not found: ID does not exist" containerID="04e09f0957bbe091d0a4bdb9e6ea0ce46e8001ce1b20b6c48a4612c0bdfbf4ca" Dec 09 10:50:48 crc kubenswrapper[4824]: I1209 10:50:48.387611 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04e09f0957bbe091d0a4bdb9e6ea0ce46e8001ce1b20b6c48a4612c0bdfbf4ca"} err="failed to get container status \"04e09f0957bbe091d0a4bdb9e6ea0ce46e8001ce1b20b6c48a4612c0bdfbf4ca\": rpc error: code = NotFound desc = could not find container \"04e09f0957bbe091d0a4bdb9e6ea0ce46e8001ce1b20b6c48a4612c0bdfbf4ca\": container with ID starting with 04e09f0957bbe091d0a4bdb9e6ea0ce46e8001ce1b20b6c48a4612c0bdfbf4ca not found: ID does not exist" Dec 09 10:50:48 crc kubenswrapper[4824]: I1209 10:50:48.387633 4824 scope.go:117] "RemoveContainer" containerID="95a5f64697925ae7ef4bf6a2eeabee3f48855304f7f56c47a59fa46d369e1a69" Dec 09 10:50:48 crc kubenswrapper[4824]: E1209 10:50:48.388102 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95a5f64697925ae7ef4bf6a2eeabee3f48855304f7f56c47a59fa46d369e1a69\": container with ID starting with 95a5f64697925ae7ef4bf6a2eeabee3f48855304f7f56c47a59fa46d369e1a69 not found: ID does not exist" containerID="95a5f64697925ae7ef4bf6a2eeabee3f48855304f7f56c47a59fa46d369e1a69" Dec 09 10:50:48 crc kubenswrapper[4824]: I1209 10:50:48.388162 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95a5f64697925ae7ef4bf6a2eeabee3f48855304f7f56c47a59fa46d369e1a69"} err="failed to get container status \"95a5f64697925ae7ef4bf6a2eeabee3f48855304f7f56c47a59fa46d369e1a69\": rpc error: code = NotFound desc = could not find container \"95a5f64697925ae7ef4bf6a2eeabee3f48855304f7f56c47a59fa46d369e1a69\": container with ID starting with 95a5f64697925ae7ef4bf6a2eeabee3f48855304f7f56c47a59fa46d369e1a69 not found: ID does not exist" Dec 09 10:50:48 crc kubenswrapper[4824]: I1209 10:50:48.388202 4824 scope.go:117] "RemoveContainer" containerID="69a07c36ebf55bfc33838e5dbf680813a0350b8327a88cb9583e1faacb9ba200" Dec 09 10:50:48 crc kubenswrapper[4824]: E1209 10:50:48.388582 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69a07c36ebf55bfc33838e5dbf680813a0350b8327a88cb9583e1faacb9ba200\": container with ID starting with 69a07c36ebf55bfc33838e5dbf680813a0350b8327a88cb9583e1faacb9ba200 not found: ID does not exist" containerID="69a07c36ebf55bfc33838e5dbf680813a0350b8327a88cb9583e1faacb9ba200" Dec 09 10:50:48 crc kubenswrapper[4824]: I1209 10:50:48.388623 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69a07c36ebf55bfc33838e5dbf680813a0350b8327a88cb9583e1faacb9ba200"} err="failed to get container status \"69a07c36ebf55bfc33838e5dbf680813a0350b8327a88cb9583e1faacb9ba200\": rpc error: code = NotFound desc = could not find container \"69a07c36ebf55bfc33838e5dbf680813a0350b8327a88cb9583e1faacb9ba200\": container with ID starting with 69a07c36ebf55bfc33838e5dbf680813a0350b8327a88cb9583e1faacb9ba200 not found: ID does not exist" Dec 09 10:50:49 crc kubenswrapper[4824]: I1209 10:50:49.924632 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ab7caf8-ce36-4db9-bc21-7d33520b96d1" path="/var/lib/kubelet/pods/2ab7caf8-ce36-4db9-bc21-7d33520b96d1/volumes" Dec 09 10:50:57 crc kubenswrapper[4824]: I1209 10:50:57.611303 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pr5tp"] Dec 09 10:50:57 crc kubenswrapper[4824]: E1209 10:50:57.612305 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ab7caf8-ce36-4db9-bc21-7d33520b96d1" containerName="extract-utilities" Dec 09 10:50:57 crc kubenswrapper[4824]: I1209 10:50:57.612320 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ab7caf8-ce36-4db9-bc21-7d33520b96d1" containerName="extract-utilities" Dec 09 10:50:57 crc kubenswrapper[4824]: E1209 10:50:57.612331 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ab7caf8-ce36-4db9-bc21-7d33520b96d1" containerName="extract-content" Dec 09 10:50:57 crc kubenswrapper[4824]: I1209 10:50:57.612337 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ab7caf8-ce36-4db9-bc21-7d33520b96d1" containerName="extract-content" Dec 09 10:50:57 crc kubenswrapper[4824]: E1209 10:50:57.612420 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ab7caf8-ce36-4db9-bc21-7d33520b96d1" containerName="registry-server" Dec 09 10:50:57 crc kubenswrapper[4824]: I1209 10:50:57.612428 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ab7caf8-ce36-4db9-bc21-7d33520b96d1" containerName="registry-server" Dec 09 10:50:57 crc kubenswrapper[4824]: I1209 10:50:57.612667 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ab7caf8-ce36-4db9-bc21-7d33520b96d1" containerName="registry-server" Dec 09 10:50:57 crc kubenswrapper[4824]: I1209 10:50:57.614951 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pr5tp" Dec 09 10:50:57 crc kubenswrapper[4824]: I1209 10:50:57.631702 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pr5tp"] Dec 09 10:50:57 crc kubenswrapper[4824]: I1209 10:50:57.735582 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgxn4\" (UniqueName: \"kubernetes.io/projected/e0cb7e3c-c501-452a-9c95-86b4468b8474-kube-api-access-wgxn4\") pod \"certified-operators-pr5tp\" (UID: \"e0cb7e3c-c501-452a-9c95-86b4468b8474\") " pod="openshift-marketplace/certified-operators-pr5tp" Dec 09 10:50:57 crc kubenswrapper[4824]: I1209 10:50:57.735954 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0cb7e3c-c501-452a-9c95-86b4468b8474-catalog-content\") pod \"certified-operators-pr5tp\" (UID: \"e0cb7e3c-c501-452a-9c95-86b4468b8474\") " pod="openshift-marketplace/certified-operators-pr5tp" Dec 09 10:50:57 crc kubenswrapper[4824]: I1209 10:50:57.736457 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0cb7e3c-c501-452a-9c95-86b4468b8474-utilities\") pod \"certified-operators-pr5tp\" (UID: \"e0cb7e3c-c501-452a-9c95-86b4468b8474\") " pod="openshift-marketplace/certified-operators-pr5tp" Dec 09 10:50:57 crc kubenswrapper[4824]: I1209 10:50:57.838920 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0cb7e3c-c501-452a-9c95-86b4468b8474-catalog-content\") pod \"certified-operators-pr5tp\" (UID: \"e0cb7e3c-c501-452a-9c95-86b4468b8474\") " pod="openshift-marketplace/certified-operators-pr5tp" Dec 09 10:50:57 crc kubenswrapper[4824]: I1209 10:50:57.839104 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0cb7e3c-c501-452a-9c95-86b4468b8474-utilities\") pod \"certified-operators-pr5tp\" (UID: \"e0cb7e3c-c501-452a-9c95-86b4468b8474\") " pod="openshift-marketplace/certified-operators-pr5tp" Dec 09 10:50:57 crc kubenswrapper[4824]: I1209 10:50:57.839201 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgxn4\" (UniqueName: \"kubernetes.io/projected/e0cb7e3c-c501-452a-9c95-86b4468b8474-kube-api-access-wgxn4\") pod \"certified-operators-pr5tp\" (UID: \"e0cb7e3c-c501-452a-9c95-86b4468b8474\") " pod="openshift-marketplace/certified-operators-pr5tp" Dec 09 10:50:57 crc kubenswrapper[4824]: I1209 10:50:57.839926 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0cb7e3c-c501-452a-9c95-86b4468b8474-catalog-content\") pod \"certified-operators-pr5tp\" (UID: \"e0cb7e3c-c501-452a-9c95-86b4468b8474\") " pod="openshift-marketplace/certified-operators-pr5tp" Dec 09 10:50:57 crc kubenswrapper[4824]: I1209 10:50:57.840029 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0cb7e3c-c501-452a-9c95-86b4468b8474-utilities\") pod \"certified-operators-pr5tp\" (UID: \"e0cb7e3c-c501-452a-9c95-86b4468b8474\") " pod="openshift-marketplace/certified-operators-pr5tp" Dec 09 10:50:57 crc kubenswrapper[4824]: I1209 10:50:57.874660 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgxn4\" (UniqueName: \"kubernetes.io/projected/e0cb7e3c-c501-452a-9c95-86b4468b8474-kube-api-access-wgxn4\") pod \"certified-operators-pr5tp\" (UID: \"e0cb7e3c-c501-452a-9c95-86b4468b8474\") " pod="openshift-marketplace/certified-operators-pr5tp" Dec 09 10:50:57 crc kubenswrapper[4824]: I1209 10:50:57.953069 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pr5tp" Dec 09 10:50:58 crc kubenswrapper[4824]: I1209 10:50:58.532691 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pr5tp"] Dec 09 10:50:59 crc kubenswrapper[4824]: I1209 10:50:59.424439 4824 generic.go:334] "Generic (PLEG): container finished" podID="e0cb7e3c-c501-452a-9c95-86b4468b8474" containerID="06b14aa6482d1bf625baf71f5de147cd4004445afb30ba600a504f4ec0262c69" exitCode=0 Dec 09 10:50:59 crc kubenswrapper[4824]: I1209 10:50:59.424636 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pr5tp" event={"ID":"e0cb7e3c-c501-452a-9c95-86b4468b8474","Type":"ContainerDied","Data":"06b14aa6482d1bf625baf71f5de147cd4004445afb30ba600a504f4ec0262c69"} Dec 09 10:50:59 crc kubenswrapper[4824]: I1209 10:50:59.424746 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pr5tp" event={"ID":"e0cb7e3c-c501-452a-9c95-86b4468b8474","Type":"ContainerStarted","Data":"a469a5622490130cbe1b13641e4e30cfcd60cb17917e67d17d9d2681553f6565"} Dec 09 10:51:00 crc kubenswrapper[4824]: I1209 10:51:00.442922 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pr5tp" event={"ID":"e0cb7e3c-c501-452a-9c95-86b4468b8474","Type":"ContainerStarted","Data":"7549fee0f1dbdb9ac7f298b46cbe671f6fb171b949c94c48da70875466f58f2b"} Dec 09 10:51:02 crc kubenswrapper[4824]: I1209 10:51:02.465249 4824 generic.go:334] "Generic (PLEG): container finished" podID="e0cb7e3c-c501-452a-9c95-86b4468b8474" containerID="7549fee0f1dbdb9ac7f298b46cbe671f6fb171b949c94c48da70875466f58f2b" exitCode=0 Dec 09 10:51:02 crc kubenswrapper[4824]: I1209 10:51:02.465729 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pr5tp" event={"ID":"e0cb7e3c-c501-452a-9c95-86b4468b8474","Type":"ContainerDied","Data":"7549fee0f1dbdb9ac7f298b46cbe671f6fb171b949c94c48da70875466f58f2b"} Dec 09 10:51:02 crc kubenswrapper[4824]: I1209 10:51:02.860689 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:51:02 crc kubenswrapper[4824]: I1209 10:51:02.860807 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:51:03 crc kubenswrapper[4824]: I1209 10:51:03.484364 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pr5tp" event={"ID":"e0cb7e3c-c501-452a-9c95-86b4468b8474","Type":"ContainerStarted","Data":"da0c834d92ae3ef43053a9c51e6467428d7669bba362d9d269fae6395f55650c"} Dec 09 10:51:03 crc kubenswrapper[4824]: I1209 10:51:03.512941 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pr5tp" podStartSLOduration=3.094275521 podStartE2EDuration="6.512915844s" podCreationTimestamp="2025-12-09 10:50:57 +0000 UTC" firstStartedPulling="2025-12-09 10:50:59.428269203 +0000 UTC m=+3815.762773870" lastFinishedPulling="2025-12-09 10:51:02.846909526 +0000 UTC m=+3819.181414193" observedRunningTime="2025-12-09 10:51:03.507299407 +0000 UTC m=+3819.841804084" watchObservedRunningTime="2025-12-09 10:51:03.512915844 +0000 UTC m=+3819.847420511" Dec 09 10:51:07 crc kubenswrapper[4824]: I1209 10:51:07.953657 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pr5tp" Dec 09 10:51:07 crc kubenswrapper[4824]: I1209 10:51:07.954020 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pr5tp" Dec 09 10:51:08 crc kubenswrapper[4824]: I1209 10:51:08.005706 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pr5tp" Dec 09 10:51:08 crc kubenswrapper[4824]: I1209 10:51:08.597102 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pr5tp" Dec 09 10:51:08 crc kubenswrapper[4824]: I1209 10:51:08.651169 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pr5tp"] Dec 09 10:51:10 crc kubenswrapper[4824]: I1209 10:51:10.556754 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pr5tp" podUID="e0cb7e3c-c501-452a-9c95-86b4468b8474" containerName="registry-server" containerID="cri-o://da0c834d92ae3ef43053a9c51e6467428d7669bba362d9d269fae6395f55650c" gracePeriod=2 Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.101661 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pr5tp" Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.110793 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0cb7e3c-c501-452a-9c95-86b4468b8474-catalog-content\") pod \"e0cb7e3c-c501-452a-9c95-86b4468b8474\" (UID: \"e0cb7e3c-c501-452a-9c95-86b4468b8474\") " Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.110861 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgxn4\" (UniqueName: \"kubernetes.io/projected/e0cb7e3c-c501-452a-9c95-86b4468b8474-kube-api-access-wgxn4\") pod \"e0cb7e3c-c501-452a-9c95-86b4468b8474\" (UID: \"e0cb7e3c-c501-452a-9c95-86b4468b8474\") " Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.110936 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0cb7e3c-c501-452a-9c95-86b4468b8474-utilities\") pod \"e0cb7e3c-c501-452a-9c95-86b4468b8474\" (UID: \"e0cb7e3c-c501-452a-9c95-86b4468b8474\") " Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.111857 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0cb7e3c-c501-452a-9c95-86b4468b8474-utilities" (OuterVolumeSpecName: "utilities") pod "e0cb7e3c-c501-452a-9c95-86b4468b8474" (UID: "e0cb7e3c-c501-452a-9c95-86b4468b8474"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.117100 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0cb7e3c-c501-452a-9c95-86b4468b8474-kube-api-access-wgxn4" (OuterVolumeSpecName: "kube-api-access-wgxn4") pod "e0cb7e3c-c501-452a-9c95-86b4468b8474" (UID: "e0cb7e3c-c501-452a-9c95-86b4468b8474"). InnerVolumeSpecName "kube-api-access-wgxn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.191663 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0cb7e3c-c501-452a-9c95-86b4468b8474-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e0cb7e3c-c501-452a-9c95-86b4468b8474" (UID: "e0cb7e3c-c501-452a-9c95-86b4468b8474"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.213758 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0cb7e3c-c501-452a-9c95-86b4468b8474-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.213916 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0cb7e3c-c501-452a-9c95-86b4468b8474-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.213935 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgxn4\" (UniqueName: \"kubernetes.io/projected/e0cb7e3c-c501-452a-9c95-86b4468b8474-kube-api-access-wgxn4\") on node \"crc\" DevicePath \"\"" Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.572511 4824 generic.go:334] "Generic (PLEG): container finished" podID="e0cb7e3c-c501-452a-9c95-86b4468b8474" containerID="da0c834d92ae3ef43053a9c51e6467428d7669bba362d9d269fae6395f55650c" exitCode=0 Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.572574 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pr5tp" Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.572584 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pr5tp" event={"ID":"e0cb7e3c-c501-452a-9c95-86b4468b8474","Type":"ContainerDied","Data":"da0c834d92ae3ef43053a9c51e6467428d7669bba362d9d269fae6395f55650c"} Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.572645 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pr5tp" event={"ID":"e0cb7e3c-c501-452a-9c95-86b4468b8474","Type":"ContainerDied","Data":"a469a5622490130cbe1b13641e4e30cfcd60cb17917e67d17d9d2681553f6565"} Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.572706 4824 scope.go:117] "RemoveContainer" containerID="da0c834d92ae3ef43053a9c51e6467428d7669bba362d9d269fae6395f55650c" Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.606628 4824 scope.go:117] "RemoveContainer" containerID="7549fee0f1dbdb9ac7f298b46cbe671f6fb171b949c94c48da70875466f58f2b" Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.615760 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pr5tp"] Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.630410 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pr5tp"] Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.632972 4824 scope.go:117] "RemoveContainer" containerID="06b14aa6482d1bf625baf71f5de147cd4004445afb30ba600a504f4ec0262c69" Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.693558 4824 scope.go:117] "RemoveContainer" containerID="da0c834d92ae3ef43053a9c51e6467428d7669bba362d9d269fae6395f55650c" Dec 09 10:51:11 crc kubenswrapper[4824]: E1209 10:51:11.693995 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da0c834d92ae3ef43053a9c51e6467428d7669bba362d9d269fae6395f55650c\": container with ID starting with da0c834d92ae3ef43053a9c51e6467428d7669bba362d9d269fae6395f55650c not found: ID does not exist" containerID="da0c834d92ae3ef43053a9c51e6467428d7669bba362d9d269fae6395f55650c" Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.694032 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da0c834d92ae3ef43053a9c51e6467428d7669bba362d9d269fae6395f55650c"} err="failed to get container status \"da0c834d92ae3ef43053a9c51e6467428d7669bba362d9d269fae6395f55650c\": rpc error: code = NotFound desc = could not find container \"da0c834d92ae3ef43053a9c51e6467428d7669bba362d9d269fae6395f55650c\": container with ID starting with da0c834d92ae3ef43053a9c51e6467428d7669bba362d9d269fae6395f55650c not found: ID does not exist" Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.694077 4824 scope.go:117] "RemoveContainer" containerID="7549fee0f1dbdb9ac7f298b46cbe671f6fb171b949c94c48da70875466f58f2b" Dec 09 10:51:11 crc kubenswrapper[4824]: E1209 10:51:11.694379 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7549fee0f1dbdb9ac7f298b46cbe671f6fb171b949c94c48da70875466f58f2b\": container with ID starting with 7549fee0f1dbdb9ac7f298b46cbe671f6fb171b949c94c48da70875466f58f2b not found: ID does not exist" containerID="7549fee0f1dbdb9ac7f298b46cbe671f6fb171b949c94c48da70875466f58f2b" Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.694408 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7549fee0f1dbdb9ac7f298b46cbe671f6fb171b949c94c48da70875466f58f2b"} err="failed to get container status \"7549fee0f1dbdb9ac7f298b46cbe671f6fb171b949c94c48da70875466f58f2b\": rpc error: code = NotFound desc = could not find container \"7549fee0f1dbdb9ac7f298b46cbe671f6fb171b949c94c48da70875466f58f2b\": container with ID starting with 7549fee0f1dbdb9ac7f298b46cbe671f6fb171b949c94c48da70875466f58f2b not found: ID does not exist" Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.694428 4824 scope.go:117] "RemoveContainer" containerID="06b14aa6482d1bf625baf71f5de147cd4004445afb30ba600a504f4ec0262c69" Dec 09 10:51:11 crc kubenswrapper[4824]: E1209 10:51:11.694662 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06b14aa6482d1bf625baf71f5de147cd4004445afb30ba600a504f4ec0262c69\": container with ID starting with 06b14aa6482d1bf625baf71f5de147cd4004445afb30ba600a504f4ec0262c69 not found: ID does not exist" containerID="06b14aa6482d1bf625baf71f5de147cd4004445afb30ba600a504f4ec0262c69" Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.694702 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06b14aa6482d1bf625baf71f5de147cd4004445afb30ba600a504f4ec0262c69"} err="failed to get container status \"06b14aa6482d1bf625baf71f5de147cd4004445afb30ba600a504f4ec0262c69\": rpc error: code = NotFound desc = could not find container \"06b14aa6482d1bf625baf71f5de147cd4004445afb30ba600a504f4ec0262c69\": container with ID starting with 06b14aa6482d1bf625baf71f5de147cd4004445afb30ba600a504f4ec0262c69 not found: ID does not exist" Dec 09 10:51:11 crc kubenswrapper[4824]: I1209 10:51:11.930246 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0cb7e3c-c501-452a-9c95-86b4468b8474" path="/var/lib/kubelet/pods/e0cb7e3c-c501-452a-9c95-86b4468b8474/volumes" Dec 09 10:51:32 crc kubenswrapper[4824]: I1209 10:51:32.861208 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:51:32 crc kubenswrapper[4824]: I1209 10:51:32.861709 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:51:32 crc kubenswrapper[4824]: I1209 10:51:32.861757 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 10:51:32 crc kubenswrapper[4824]: I1209 10:51:32.862370 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e"} pod="openshift-machine-config-operator/machine-config-daemon-dth8x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 10:51:32 crc kubenswrapper[4824]: I1209 10:51:32.862425 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" containerID="cri-o://fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" gracePeriod=600 Dec 09 10:51:32 crc kubenswrapper[4824]: E1209 10:51:32.982967 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:51:33 crc kubenswrapper[4824]: I1209 10:51:33.844433 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" exitCode=0 Dec 09 10:51:33 crc kubenswrapper[4824]: I1209 10:51:33.844561 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerDied","Data":"fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e"} Dec 09 10:51:33 crc kubenswrapper[4824]: I1209 10:51:33.845366 4824 scope.go:117] "RemoveContainer" containerID="7050a696e11544a6263f8787d7539e141eb5adea02be70809cc2757af296657c" Dec 09 10:51:33 crc kubenswrapper[4824]: I1209 10:51:33.845564 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:51:33 crc kubenswrapper[4824]: E1209 10:51:33.846512 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:51:34 crc kubenswrapper[4824]: I1209 10:51:34.859369 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:51:34 crc kubenswrapper[4824]: E1209 10:51:34.859941 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:51:46 crc kubenswrapper[4824]: I1209 10:51:46.911467 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:51:46 crc kubenswrapper[4824]: E1209 10:51:46.914192 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:51:58 crc kubenswrapper[4824]: I1209 10:51:58.911178 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:51:58 crc kubenswrapper[4824]: E1209 10:51:58.912167 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:52:09 crc kubenswrapper[4824]: I1209 10:52:09.929532 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:52:09 crc kubenswrapper[4824]: E1209 10:52:09.930363 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:52:24 crc kubenswrapper[4824]: I1209 10:52:24.910505 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:52:24 crc kubenswrapper[4824]: E1209 10:52:24.911527 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:52:37 crc kubenswrapper[4824]: I1209 10:52:37.911294 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:52:37 crc kubenswrapper[4824]: E1209 10:52:37.912084 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:52:48 crc kubenswrapper[4824]: I1209 10:52:48.912012 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:52:48 crc kubenswrapper[4824]: E1209 10:52:48.912704 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:53:00 crc kubenswrapper[4824]: I1209 10:53:00.910836 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:53:00 crc kubenswrapper[4824]: E1209 10:53:00.911953 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:53:15 crc kubenswrapper[4824]: I1209 10:53:15.911290 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:53:15 crc kubenswrapper[4824]: E1209 10:53:15.912274 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:53:27 crc kubenswrapper[4824]: I1209 10:53:27.912252 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:53:27 crc kubenswrapper[4824]: E1209 10:53:27.916986 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:53:38 crc kubenswrapper[4824]: I1209 10:53:38.911461 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:53:38 crc kubenswrapper[4824]: E1209 10:53:38.913940 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:53:50 crc kubenswrapper[4824]: I1209 10:53:50.911366 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:53:50 crc kubenswrapper[4824]: E1209 10:53:50.912252 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:54:02 crc kubenswrapper[4824]: I1209 10:54:02.911760 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:54:02 crc kubenswrapper[4824]: E1209 10:54:02.912582 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:54:13 crc kubenswrapper[4824]: I1209 10:54:13.918256 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:54:13 crc kubenswrapper[4824]: E1209 10:54:13.919304 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:54:27 crc kubenswrapper[4824]: I1209 10:54:27.911330 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:54:27 crc kubenswrapper[4824]: E1209 10:54:27.913238 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:54:42 crc kubenswrapper[4824]: I1209 10:54:42.911977 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:54:42 crc kubenswrapper[4824]: E1209 10:54:42.913190 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:54:44 crc kubenswrapper[4824]: I1209 10:54:44.330311 4824 trace.go:236] Trace[1657111023]: "Calculate volume metrics of persistence for pod openstack/rabbitmq-cell1-server-0" (09-Dec-2025 10:54:41.278) (total time: 3051ms): Dec 09 10:54:44 crc kubenswrapper[4824]: Trace[1657111023]: [3.051271584s] [3.051271584s] END Dec 09 10:54:53 crc kubenswrapper[4824]: I1209 10:54:53.925605 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:54:53 crc kubenswrapper[4824]: E1209 10:54:53.926434 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:54:56 crc kubenswrapper[4824]: I1209 10:54:56.069288 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-krzfq"] Dec 09 10:54:56 crc kubenswrapper[4824]: E1209 10:54:56.070482 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0cb7e3c-c501-452a-9c95-86b4468b8474" containerName="extract-content" Dec 09 10:54:56 crc kubenswrapper[4824]: I1209 10:54:56.070496 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0cb7e3c-c501-452a-9c95-86b4468b8474" containerName="extract-content" Dec 09 10:54:56 crc kubenswrapper[4824]: E1209 10:54:56.070509 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0cb7e3c-c501-452a-9c95-86b4468b8474" containerName="extract-utilities" Dec 09 10:54:56 crc kubenswrapper[4824]: I1209 10:54:56.070516 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0cb7e3c-c501-452a-9c95-86b4468b8474" containerName="extract-utilities" Dec 09 10:54:56 crc kubenswrapper[4824]: E1209 10:54:56.070533 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0cb7e3c-c501-452a-9c95-86b4468b8474" containerName="registry-server" Dec 09 10:54:56 crc kubenswrapper[4824]: I1209 10:54:56.070540 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0cb7e3c-c501-452a-9c95-86b4468b8474" containerName="registry-server" Dec 09 10:54:56 crc kubenswrapper[4824]: I1209 10:54:56.070770 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0cb7e3c-c501-452a-9c95-86b4468b8474" containerName="registry-server" Dec 09 10:54:56 crc kubenswrapper[4824]: I1209 10:54:56.085110 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-krzfq"] Dec 09 10:54:56 crc kubenswrapper[4824]: I1209 10:54:56.085453 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krzfq" Dec 09 10:54:56 crc kubenswrapper[4824]: I1209 10:54:56.210892 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c483758-75b9-4401-9d8d-6dd7d5fed7fe-catalog-content\") pod \"redhat-operators-krzfq\" (UID: \"0c483758-75b9-4401-9d8d-6dd7d5fed7fe\") " pod="openshift-marketplace/redhat-operators-krzfq" Dec 09 10:54:56 crc kubenswrapper[4824]: I1209 10:54:56.211544 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c483758-75b9-4401-9d8d-6dd7d5fed7fe-utilities\") pod \"redhat-operators-krzfq\" (UID: \"0c483758-75b9-4401-9d8d-6dd7d5fed7fe\") " pod="openshift-marketplace/redhat-operators-krzfq" Dec 09 10:54:56 crc kubenswrapper[4824]: I1209 10:54:56.212686 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm4v4\" (UniqueName: \"kubernetes.io/projected/0c483758-75b9-4401-9d8d-6dd7d5fed7fe-kube-api-access-tm4v4\") pod \"redhat-operators-krzfq\" (UID: \"0c483758-75b9-4401-9d8d-6dd7d5fed7fe\") " pod="openshift-marketplace/redhat-operators-krzfq" Dec 09 10:54:56 crc kubenswrapper[4824]: I1209 10:54:56.315344 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm4v4\" (UniqueName: \"kubernetes.io/projected/0c483758-75b9-4401-9d8d-6dd7d5fed7fe-kube-api-access-tm4v4\") pod \"redhat-operators-krzfq\" (UID: \"0c483758-75b9-4401-9d8d-6dd7d5fed7fe\") " pod="openshift-marketplace/redhat-operators-krzfq" Dec 09 10:54:56 crc kubenswrapper[4824]: I1209 10:54:56.315515 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c483758-75b9-4401-9d8d-6dd7d5fed7fe-catalog-content\") pod \"redhat-operators-krzfq\" (UID: \"0c483758-75b9-4401-9d8d-6dd7d5fed7fe\") " pod="openshift-marketplace/redhat-operators-krzfq" Dec 09 10:54:56 crc kubenswrapper[4824]: I1209 10:54:56.315565 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c483758-75b9-4401-9d8d-6dd7d5fed7fe-utilities\") pod \"redhat-operators-krzfq\" (UID: \"0c483758-75b9-4401-9d8d-6dd7d5fed7fe\") " pod="openshift-marketplace/redhat-operators-krzfq" Dec 09 10:54:56 crc kubenswrapper[4824]: I1209 10:54:56.316021 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c483758-75b9-4401-9d8d-6dd7d5fed7fe-utilities\") pod \"redhat-operators-krzfq\" (UID: \"0c483758-75b9-4401-9d8d-6dd7d5fed7fe\") " pod="openshift-marketplace/redhat-operators-krzfq" Dec 09 10:54:56 crc kubenswrapper[4824]: I1209 10:54:56.316125 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c483758-75b9-4401-9d8d-6dd7d5fed7fe-catalog-content\") pod \"redhat-operators-krzfq\" (UID: \"0c483758-75b9-4401-9d8d-6dd7d5fed7fe\") " pod="openshift-marketplace/redhat-operators-krzfq" Dec 09 10:54:56 crc kubenswrapper[4824]: I1209 10:54:56.748644 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm4v4\" (UniqueName: \"kubernetes.io/projected/0c483758-75b9-4401-9d8d-6dd7d5fed7fe-kube-api-access-tm4v4\") pod \"redhat-operators-krzfq\" (UID: \"0c483758-75b9-4401-9d8d-6dd7d5fed7fe\") " pod="openshift-marketplace/redhat-operators-krzfq" Dec 09 10:54:57 crc kubenswrapper[4824]: I1209 10:54:57.015349 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krzfq" Dec 09 10:54:57 crc kubenswrapper[4824]: W1209 10:54:57.524690 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c483758_75b9_4401_9d8d_6dd7d5fed7fe.slice/crio-5595adaf880c9f71c282ddb17724b16f23d82a9976b0498c65f6ab36b92c4150 WatchSource:0}: Error finding container 5595adaf880c9f71c282ddb17724b16f23d82a9976b0498c65f6ab36b92c4150: Status 404 returned error can't find the container with id 5595adaf880c9f71c282ddb17724b16f23d82a9976b0498c65f6ab36b92c4150 Dec 09 10:54:57 crc kubenswrapper[4824]: I1209 10:54:57.524982 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-krzfq"] Dec 09 10:54:57 crc kubenswrapper[4824]: I1209 10:54:57.613791 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krzfq" event={"ID":"0c483758-75b9-4401-9d8d-6dd7d5fed7fe","Type":"ContainerStarted","Data":"5595adaf880c9f71c282ddb17724b16f23d82a9976b0498c65f6ab36b92c4150"} Dec 09 10:54:58 crc kubenswrapper[4824]: I1209 10:54:58.629061 4824 generic.go:334] "Generic (PLEG): container finished" podID="0c483758-75b9-4401-9d8d-6dd7d5fed7fe" containerID="404c2401de356816d5be22c93b6e7c9f3939aebb4f85e78d05dfea8ad61f0031" exitCode=0 Dec 09 10:54:58 crc kubenswrapper[4824]: I1209 10:54:58.629147 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krzfq" event={"ID":"0c483758-75b9-4401-9d8d-6dd7d5fed7fe","Type":"ContainerDied","Data":"404c2401de356816d5be22c93b6e7c9f3939aebb4f85e78d05dfea8ad61f0031"} Dec 09 10:54:59 crc kubenswrapper[4824]: I1209 10:54:59.642587 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krzfq" event={"ID":"0c483758-75b9-4401-9d8d-6dd7d5fed7fe","Type":"ContainerStarted","Data":"75617c9d4f6cf73e2adf92176ed917f6a3c71364e8c46b336b417fdf32732d26"} Dec 09 10:55:04 crc kubenswrapper[4824]: I1209 10:55:04.705138 4824 generic.go:334] "Generic (PLEG): container finished" podID="0c483758-75b9-4401-9d8d-6dd7d5fed7fe" containerID="75617c9d4f6cf73e2adf92176ed917f6a3c71364e8c46b336b417fdf32732d26" exitCode=0 Dec 09 10:55:04 crc kubenswrapper[4824]: I1209 10:55:04.705187 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krzfq" event={"ID":"0c483758-75b9-4401-9d8d-6dd7d5fed7fe","Type":"ContainerDied","Data":"75617c9d4f6cf73e2adf92176ed917f6a3c71364e8c46b336b417fdf32732d26"} Dec 09 10:55:05 crc kubenswrapper[4824]: I1209 10:55:05.725080 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krzfq" event={"ID":"0c483758-75b9-4401-9d8d-6dd7d5fed7fe","Type":"ContainerStarted","Data":"001ebedcb1e2f242acefff674594cf2efde40ecb667efa7306486b3d1501a318"} Dec 09 10:55:05 crc kubenswrapper[4824]: I1209 10:55:05.755045 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-krzfq" podStartSLOduration=3.232358972 podStartE2EDuration="9.754988451s" podCreationTimestamp="2025-12-09 10:54:56 +0000 UTC" firstStartedPulling="2025-12-09 10:54:58.633643709 +0000 UTC m=+4054.968148386" lastFinishedPulling="2025-12-09 10:55:05.156273188 +0000 UTC m=+4061.490777865" observedRunningTime="2025-12-09 10:55:05.750276973 +0000 UTC m=+4062.084781700" watchObservedRunningTime="2025-12-09 10:55:05.754988451 +0000 UTC m=+4062.089493128" Dec 09 10:55:07 crc kubenswrapper[4824]: I1209 10:55:07.101740 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-krzfq" Dec 09 10:55:07 crc kubenswrapper[4824]: I1209 10:55:07.102413 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-krzfq" Dec 09 10:55:07 crc kubenswrapper[4824]: I1209 10:55:07.912458 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:55:07 crc kubenswrapper[4824]: E1209 10:55:07.913087 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:55:08 crc kubenswrapper[4824]: I1209 10:55:08.152512 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-krzfq" podUID="0c483758-75b9-4401-9d8d-6dd7d5fed7fe" containerName="registry-server" probeResult="failure" output=< Dec 09 10:55:08 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 10:55:08 crc kubenswrapper[4824]: > Dec 09 10:55:17 crc kubenswrapper[4824]: I1209 10:55:17.069967 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-krzfq" Dec 09 10:55:17 crc kubenswrapper[4824]: I1209 10:55:17.128236 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-krzfq" Dec 09 10:55:17 crc kubenswrapper[4824]: I1209 10:55:17.320073 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-krzfq"] Dec 09 10:55:18 crc kubenswrapper[4824]: I1209 10:55:18.993122 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-krzfq" podUID="0c483758-75b9-4401-9d8d-6dd7d5fed7fe" containerName="registry-server" containerID="cri-o://001ebedcb1e2f242acefff674594cf2efde40ecb667efa7306486b3d1501a318" gracePeriod=2 Dec 09 10:55:19 crc kubenswrapper[4824]: I1209 10:55:19.526623 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krzfq" Dec 09 10:55:19 crc kubenswrapper[4824]: I1209 10:55:19.579987 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c483758-75b9-4401-9d8d-6dd7d5fed7fe-catalog-content\") pod \"0c483758-75b9-4401-9d8d-6dd7d5fed7fe\" (UID: \"0c483758-75b9-4401-9d8d-6dd7d5fed7fe\") " Dec 09 10:55:19 crc kubenswrapper[4824]: I1209 10:55:19.580224 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tm4v4\" (UniqueName: \"kubernetes.io/projected/0c483758-75b9-4401-9d8d-6dd7d5fed7fe-kube-api-access-tm4v4\") pod \"0c483758-75b9-4401-9d8d-6dd7d5fed7fe\" (UID: \"0c483758-75b9-4401-9d8d-6dd7d5fed7fe\") " Dec 09 10:55:19 crc kubenswrapper[4824]: I1209 10:55:19.580340 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c483758-75b9-4401-9d8d-6dd7d5fed7fe-utilities\") pod \"0c483758-75b9-4401-9d8d-6dd7d5fed7fe\" (UID: \"0c483758-75b9-4401-9d8d-6dd7d5fed7fe\") " Dec 09 10:55:19 crc kubenswrapper[4824]: I1209 10:55:19.581212 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c483758-75b9-4401-9d8d-6dd7d5fed7fe-utilities" (OuterVolumeSpecName: "utilities") pod "0c483758-75b9-4401-9d8d-6dd7d5fed7fe" (UID: "0c483758-75b9-4401-9d8d-6dd7d5fed7fe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:55:19 crc kubenswrapper[4824]: I1209 10:55:19.642638 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c483758-75b9-4401-9d8d-6dd7d5fed7fe-kube-api-access-tm4v4" (OuterVolumeSpecName: "kube-api-access-tm4v4") pod "0c483758-75b9-4401-9d8d-6dd7d5fed7fe" (UID: "0c483758-75b9-4401-9d8d-6dd7d5fed7fe"). InnerVolumeSpecName "kube-api-access-tm4v4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:55:19 crc kubenswrapper[4824]: I1209 10:55:19.683320 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tm4v4\" (UniqueName: \"kubernetes.io/projected/0c483758-75b9-4401-9d8d-6dd7d5fed7fe-kube-api-access-tm4v4\") on node \"crc\" DevicePath \"\"" Dec 09 10:55:19 crc kubenswrapper[4824]: I1209 10:55:19.683365 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c483758-75b9-4401-9d8d-6dd7d5fed7fe-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 10:55:19 crc kubenswrapper[4824]: I1209 10:55:19.716324 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c483758-75b9-4401-9d8d-6dd7d5fed7fe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c483758-75b9-4401-9d8d-6dd7d5fed7fe" (UID: "0c483758-75b9-4401-9d8d-6dd7d5fed7fe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:55:19 crc kubenswrapper[4824]: I1209 10:55:19.785981 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c483758-75b9-4401-9d8d-6dd7d5fed7fe-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 10:55:20 crc kubenswrapper[4824]: I1209 10:55:20.096044 4824 generic.go:334] "Generic (PLEG): container finished" podID="0c483758-75b9-4401-9d8d-6dd7d5fed7fe" containerID="001ebedcb1e2f242acefff674594cf2efde40ecb667efa7306486b3d1501a318" exitCode=0 Dec 09 10:55:20 crc kubenswrapper[4824]: I1209 10:55:20.096096 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krzfq" event={"ID":"0c483758-75b9-4401-9d8d-6dd7d5fed7fe","Type":"ContainerDied","Data":"001ebedcb1e2f242acefff674594cf2efde40ecb667efa7306486b3d1501a318"} Dec 09 10:55:20 crc kubenswrapper[4824]: I1209 10:55:20.096127 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krzfq" Dec 09 10:55:20 crc kubenswrapper[4824]: I1209 10:55:20.096183 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krzfq" event={"ID":"0c483758-75b9-4401-9d8d-6dd7d5fed7fe","Type":"ContainerDied","Data":"5595adaf880c9f71c282ddb17724b16f23d82a9976b0498c65f6ab36b92c4150"} Dec 09 10:55:20 crc kubenswrapper[4824]: I1209 10:55:20.096212 4824 scope.go:117] "RemoveContainer" containerID="001ebedcb1e2f242acefff674594cf2efde40ecb667efa7306486b3d1501a318" Dec 09 10:55:20 crc kubenswrapper[4824]: I1209 10:55:20.126199 4824 scope.go:117] "RemoveContainer" containerID="75617c9d4f6cf73e2adf92176ed917f6a3c71364e8c46b336b417fdf32732d26" Dec 09 10:55:20 crc kubenswrapper[4824]: I1209 10:55:20.131136 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-krzfq"] Dec 09 10:55:20 crc kubenswrapper[4824]: I1209 10:55:20.142363 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-krzfq"] Dec 09 10:55:20 crc kubenswrapper[4824]: I1209 10:55:20.162365 4824 scope.go:117] "RemoveContainer" containerID="404c2401de356816d5be22c93b6e7c9f3939aebb4f85e78d05dfea8ad61f0031" Dec 09 10:55:20 crc kubenswrapper[4824]: I1209 10:55:20.215428 4824 scope.go:117] "RemoveContainer" containerID="001ebedcb1e2f242acefff674594cf2efde40ecb667efa7306486b3d1501a318" Dec 09 10:55:20 crc kubenswrapper[4824]: E1209 10:55:20.215943 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"001ebedcb1e2f242acefff674594cf2efde40ecb667efa7306486b3d1501a318\": container with ID starting with 001ebedcb1e2f242acefff674594cf2efde40ecb667efa7306486b3d1501a318 not found: ID does not exist" containerID="001ebedcb1e2f242acefff674594cf2efde40ecb667efa7306486b3d1501a318" Dec 09 10:55:20 crc kubenswrapper[4824]: I1209 10:55:20.216010 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"001ebedcb1e2f242acefff674594cf2efde40ecb667efa7306486b3d1501a318"} err="failed to get container status \"001ebedcb1e2f242acefff674594cf2efde40ecb667efa7306486b3d1501a318\": rpc error: code = NotFound desc = could not find container \"001ebedcb1e2f242acefff674594cf2efde40ecb667efa7306486b3d1501a318\": container with ID starting with 001ebedcb1e2f242acefff674594cf2efde40ecb667efa7306486b3d1501a318 not found: ID does not exist" Dec 09 10:55:20 crc kubenswrapper[4824]: I1209 10:55:20.216045 4824 scope.go:117] "RemoveContainer" containerID="75617c9d4f6cf73e2adf92176ed917f6a3c71364e8c46b336b417fdf32732d26" Dec 09 10:55:20 crc kubenswrapper[4824]: E1209 10:55:20.216484 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75617c9d4f6cf73e2adf92176ed917f6a3c71364e8c46b336b417fdf32732d26\": container with ID starting with 75617c9d4f6cf73e2adf92176ed917f6a3c71364e8c46b336b417fdf32732d26 not found: ID does not exist" containerID="75617c9d4f6cf73e2adf92176ed917f6a3c71364e8c46b336b417fdf32732d26" Dec 09 10:55:20 crc kubenswrapper[4824]: I1209 10:55:20.216551 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75617c9d4f6cf73e2adf92176ed917f6a3c71364e8c46b336b417fdf32732d26"} err="failed to get container status \"75617c9d4f6cf73e2adf92176ed917f6a3c71364e8c46b336b417fdf32732d26\": rpc error: code = NotFound desc = could not find container \"75617c9d4f6cf73e2adf92176ed917f6a3c71364e8c46b336b417fdf32732d26\": container with ID starting with 75617c9d4f6cf73e2adf92176ed917f6a3c71364e8c46b336b417fdf32732d26 not found: ID does not exist" Dec 09 10:55:20 crc kubenswrapper[4824]: I1209 10:55:20.216592 4824 scope.go:117] "RemoveContainer" containerID="404c2401de356816d5be22c93b6e7c9f3939aebb4f85e78d05dfea8ad61f0031" Dec 09 10:55:20 crc kubenswrapper[4824]: E1209 10:55:20.218205 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"404c2401de356816d5be22c93b6e7c9f3939aebb4f85e78d05dfea8ad61f0031\": container with ID starting with 404c2401de356816d5be22c93b6e7c9f3939aebb4f85e78d05dfea8ad61f0031 not found: ID does not exist" containerID="404c2401de356816d5be22c93b6e7c9f3939aebb4f85e78d05dfea8ad61f0031" Dec 09 10:55:20 crc kubenswrapper[4824]: I1209 10:55:20.218332 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"404c2401de356816d5be22c93b6e7c9f3939aebb4f85e78d05dfea8ad61f0031"} err="failed to get container status \"404c2401de356816d5be22c93b6e7c9f3939aebb4f85e78d05dfea8ad61f0031\": rpc error: code = NotFound desc = could not find container \"404c2401de356816d5be22c93b6e7c9f3939aebb4f85e78d05dfea8ad61f0031\": container with ID starting with 404c2401de356816d5be22c93b6e7c9f3939aebb4f85e78d05dfea8ad61f0031 not found: ID does not exist" Dec 09 10:55:21 crc kubenswrapper[4824]: I1209 10:55:21.990186 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:55:21 crc kubenswrapper[4824]: E1209 10:55:21.991128 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:55:22 crc kubenswrapper[4824]: I1209 10:55:22.037417 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c483758-75b9-4401-9d8d-6dd7d5fed7fe" path="/var/lib/kubelet/pods/0c483758-75b9-4401-9d8d-6dd7d5fed7fe/volumes" Dec 09 10:55:35 crc kubenswrapper[4824]: I1209 10:55:35.938118 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:55:35 crc kubenswrapper[4824]: E1209 10:55:35.939086 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:55:46 crc kubenswrapper[4824]: I1209 10:55:46.911315 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:55:46 crc kubenswrapper[4824]: E1209 10:55:46.912504 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:55:57 crc kubenswrapper[4824]: I1209 10:55:57.911424 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:55:57 crc kubenswrapper[4824]: E1209 10:55:57.912615 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:56:07 crc kubenswrapper[4824]: I1209 10:56:07.023304 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hkjzk"] Dec 09 10:56:07 crc kubenswrapper[4824]: E1209 10:56:07.024348 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c483758-75b9-4401-9d8d-6dd7d5fed7fe" containerName="registry-server" Dec 09 10:56:07 crc kubenswrapper[4824]: I1209 10:56:07.024363 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c483758-75b9-4401-9d8d-6dd7d5fed7fe" containerName="registry-server" Dec 09 10:56:07 crc kubenswrapper[4824]: E1209 10:56:07.024390 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c483758-75b9-4401-9d8d-6dd7d5fed7fe" containerName="extract-content" Dec 09 10:56:07 crc kubenswrapper[4824]: I1209 10:56:07.024395 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c483758-75b9-4401-9d8d-6dd7d5fed7fe" containerName="extract-content" Dec 09 10:56:07 crc kubenswrapper[4824]: E1209 10:56:07.024419 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c483758-75b9-4401-9d8d-6dd7d5fed7fe" containerName="extract-utilities" Dec 09 10:56:07 crc kubenswrapper[4824]: I1209 10:56:07.024425 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c483758-75b9-4401-9d8d-6dd7d5fed7fe" containerName="extract-utilities" Dec 09 10:56:07 crc kubenswrapper[4824]: I1209 10:56:07.024669 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c483758-75b9-4401-9d8d-6dd7d5fed7fe" containerName="registry-server" Dec 09 10:56:07 crc kubenswrapper[4824]: I1209 10:56:07.026620 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hkjzk" Dec 09 10:56:07 crc kubenswrapper[4824]: I1209 10:56:07.054395 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hkjzk"] Dec 09 10:56:07 crc kubenswrapper[4824]: I1209 10:56:07.196623 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4d5eeca-4d1e-49d7-9c59-b37618033fd3-utilities\") pod \"community-operators-hkjzk\" (UID: \"a4d5eeca-4d1e-49d7-9c59-b37618033fd3\") " pod="openshift-marketplace/community-operators-hkjzk" Dec 09 10:56:07 crc kubenswrapper[4824]: I1209 10:56:07.196765 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frzl5\" (UniqueName: \"kubernetes.io/projected/a4d5eeca-4d1e-49d7-9c59-b37618033fd3-kube-api-access-frzl5\") pod \"community-operators-hkjzk\" (UID: \"a4d5eeca-4d1e-49d7-9c59-b37618033fd3\") " pod="openshift-marketplace/community-operators-hkjzk" Dec 09 10:56:07 crc kubenswrapper[4824]: I1209 10:56:07.196880 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4d5eeca-4d1e-49d7-9c59-b37618033fd3-catalog-content\") pod \"community-operators-hkjzk\" (UID: \"a4d5eeca-4d1e-49d7-9c59-b37618033fd3\") " pod="openshift-marketplace/community-operators-hkjzk" Dec 09 10:56:07 crc kubenswrapper[4824]: I1209 10:56:07.299689 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4d5eeca-4d1e-49d7-9c59-b37618033fd3-catalog-content\") pod \"community-operators-hkjzk\" (UID: \"a4d5eeca-4d1e-49d7-9c59-b37618033fd3\") " pod="openshift-marketplace/community-operators-hkjzk" Dec 09 10:56:07 crc kubenswrapper[4824]: I1209 10:56:07.299865 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4d5eeca-4d1e-49d7-9c59-b37618033fd3-utilities\") pod \"community-operators-hkjzk\" (UID: \"a4d5eeca-4d1e-49d7-9c59-b37618033fd3\") " pod="openshift-marketplace/community-operators-hkjzk" Dec 09 10:56:07 crc kubenswrapper[4824]: I1209 10:56:07.300077 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frzl5\" (UniqueName: \"kubernetes.io/projected/a4d5eeca-4d1e-49d7-9c59-b37618033fd3-kube-api-access-frzl5\") pod \"community-operators-hkjzk\" (UID: \"a4d5eeca-4d1e-49d7-9c59-b37618033fd3\") " pod="openshift-marketplace/community-operators-hkjzk" Dec 09 10:56:07 crc kubenswrapper[4824]: I1209 10:56:07.300635 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4d5eeca-4d1e-49d7-9c59-b37618033fd3-catalog-content\") pod \"community-operators-hkjzk\" (UID: \"a4d5eeca-4d1e-49d7-9c59-b37618033fd3\") " pod="openshift-marketplace/community-operators-hkjzk" Dec 09 10:56:07 crc kubenswrapper[4824]: I1209 10:56:07.300714 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4d5eeca-4d1e-49d7-9c59-b37618033fd3-utilities\") pod \"community-operators-hkjzk\" (UID: \"a4d5eeca-4d1e-49d7-9c59-b37618033fd3\") " pod="openshift-marketplace/community-operators-hkjzk" Dec 09 10:56:07 crc kubenswrapper[4824]: I1209 10:56:07.323243 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frzl5\" (UniqueName: \"kubernetes.io/projected/a4d5eeca-4d1e-49d7-9c59-b37618033fd3-kube-api-access-frzl5\") pod \"community-operators-hkjzk\" (UID: \"a4d5eeca-4d1e-49d7-9c59-b37618033fd3\") " pod="openshift-marketplace/community-operators-hkjzk" Dec 09 10:56:07 crc kubenswrapper[4824]: I1209 10:56:07.355577 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hkjzk" Dec 09 10:56:08 crc kubenswrapper[4824]: I1209 10:56:08.106136 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hkjzk"] Dec 09 10:56:08 crc kubenswrapper[4824]: E1209 10:56:08.571273 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4d5eeca_4d1e_49d7_9c59_b37618033fd3.slice/crio-conmon-20e34433aa2f535e862c1e9c7d942ae34d44e43491d3029734fbe1aab5c7312d.scope\": RecentStats: unable to find data in memory cache]" Dec 09 10:56:09 crc kubenswrapper[4824]: I1209 10:56:09.016669 4824 generic.go:334] "Generic (PLEG): container finished" podID="a4d5eeca-4d1e-49d7-9c59-b37618033fd3" containerID="20e34433aa2f535e862c1e9c7d942ae34d44e43491d3029734fbe1aab5c7312d" exitCode=0 Dec 09 10:56:09 crc kubenswrapper[4824]: I1209 10:56:09.016713 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkjzk" event={"ID":"a4d5eeca-4d1e-49d7-9c59-b37618033fd3","Type":"ContainerDied","Data":"20e34433aa2f535e862c1e9c7d942ae34d44e43491d3029734fbe1aab5c7312d"} Dec 09 10:56:09 crc kubenswrapper[4824]: I1209 10:56:09.016737 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkjzk" event={"ID":"a4d5eeca-4d1e-49d7-9c59-b37618033fd3","Type":"ContainerStarted","Data":"78c5fc38d54c3e64ac410e339e0acac3df25d82cf3d934650775a97d8902a6f3"} Dec 09 10:56:09 crc kubenswrapper[4824]: I1209 10:56:09.018974 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 10:56:10 crc kubenswrapper[4824]: I1209 10:56:10.034716 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkjzk" event={"ID":"a4d5eeca-4d1e-49d7-9c59-b37618033fd3","Type":"ContainerStarted","Data":"686f152adecbc4a01864dd6e995dd9a27862affc114fb21409f443642d4c9c5c"} Dec 09 10:56:11 crc kubenswrapper[4824]: I1209 10:56:11.057613 4824 generic.go:334] "Generic (PLEG): container finished" podID="a4d5eeca-4d1e-49d7-9c59-b37618033fd3" containerID="686f152adecbc4a01864dd6e995dd9a27862affc114fb21409f443642d4c9c5c" exitCode=0 Dec 09 10:56:11 crc kubenswrapper[4824]: I1209 10:56:11.057971 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkjzk" event={"ID":"a4d5eeca-4d1e-49d7-9c59-b37618033fd3","Type":"ContainerDied","Data":"686f152adecbc4a01864dd6e995dd9a27862affc114fb21409f443642d4c9c5c"} Dec 09 10:56:12 crc kubenswrapper[4824]: I1209 10:56:12.075190 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkjzk" event={"ID":"a4d5eeca-4d1e-49d7-9c59-b37618033fd3","Type":"ContainerStarted","Data":"ad894e1b364c9e3f8c9a0c6037a4481953769d4414bb08453aead36c1c040e7e"} Dec 09 10:56:12 crc kubenswrapper[4824]: I1209 10:56:12.111276 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hkjzk" podStartSLOduration=3.694373923 podStartE2EDuration="6.11125457s" podCreationTimestamp="2025-12-09 10:56:06 +0000 UTC" firstStartedPulling="2025-12-09 10:56:09.018738638 +0000 UTC m=+4125.353243305" lastFinishedPulling="2025-12-09 10:56:11.435619285 +0000 UTC m=+4127.770123952" observedRunningTime="2025-12-09 10:56:12.098926792 +0000 UTC m=+4128.433431479" watchObservedRunningTime="2025-12-09 10:56:12.11125457 +0000 UTC m=+4128.445759237" Dec 09 10:56:12 crc kubenswrapper[4824]: I1209 10:56:12.916061 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:56:12 crc kubenswrapper[4824]: E1209 10:56:12.916683 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:56:17 crc kubenswrapper[4824]: I1209 10:56:17.356833 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hkjzk" Dec 09 10:56:17 crc kubenswrapper[4824]: I1209 10:56:17.357386 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hkjzk" Dec 09 10:56:17 crc kubenswrapper[4824]: I1209 10:56:17.500196 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hkjzk" Dec 09 10:56:18 crc kubenswrapper[4824]: I1209 10:56:18.212535 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hkjzk" Dec 09 10:56:18 crc kubenswrapper[4824]: I1209 10:56:18.272286 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hkjzk"] Dec 09 10:56:20 crc kubenswrapper[4824]: I1209 10:56:20.170127 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hkjzk" podUID="a4d5eeca-4d1e-49d7-9c59-b37618033fd3" containerName="registry-server" containerID="cri-o://ad894e1b364c9e3f8c9a0c6037a4481953769d4414bb08453aead36c1c040e7e" gracePeriod=2 Dec 09 10:56:21 crc kubenswrapper[4824]: I1209 10:56:21.182749 4824 generic.go:334] "Generic (PLEG): container finished" podID="a4d5eeca-4d1e-49d7-9c59-b37618033fd3" containerID="ad894e1b364c9e3f8c9a0c6037a4481953769d4414bb08453aead36c1c040e7e" exitCode=0 Dec 09 10:56:21 crc kubenswrapper[4824]: I1209 10:56:21.182822 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkjzk" event={"ID":"a4d5eeca-4d1e-49d7-9c59-b37618033fd3","Type":"ContainerDied","Data":"ad894e1b364c9e3f8c9a0c6037a4481953769d4414bb08453aead36c1c040e7e"} Dec 09 10:56:21 crc kubenswrapper[4824]: I1209 10:56:21.183120 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkjzk" event={"ID":"a4d5eeca-4d1e-49d7-9c59-b37618033fd3","Type":"ContainerDied","Data":"78c5fc38d54c3e64ac410e339e0acac3df25d82cf3d934650775a97d8902a6f3"} Dec 09 10:56:21 crc kubenswrapper[4824]: I1209 10:56:21.183138 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78c5fc38d54c3e64ac410e339e0acac3df25d82cf3d934650775a97d8902a6f3" Dec 09 10:56:21 crc kubenswrapper[4824]: I1209 10:56:21.190656 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hkjzk" Dec 09 10:56:21 crc kubenswrapper[4824]: I1209 10:56:21.390328 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4d5eeca-4d1e-49d7-9c59-b37618033fd3-catalog-content\") pod \"a4d5eeca-4d1e-49d7-9c59-b37618033fd3\" (UID: \"a4d5eeca-4d1e-49d7-9c59-b37618033fd3\") " Dec 09 10:56:21 crc kubenswrapper[4824]: I1209 10:56:21.390625 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frzl5\" (UniqueName: \"kubernetes.io/projected/a4d5eeca-4d1e-49d7-9c59-b37618033fd3-kube-api-access-frzl5\") pod \"a4d5eeca-4d1e-49d7-9c59-b37618033fd3\" (UID: \"a4d5eeca-4d1e-49d7-9c59-b37618033fd3\") " Dec 09 10:56:21 crc kubenswrapper[4824]: I1209 10:56:21.390849 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4d5eeca-4d1e-49d7-9c59-b37618033fd3-utilities\") pod \"a4d5eeca-4d1e-49d7-9c59-b37618033fd3\" (UID: \"a4d5eeca-4d1e-49d7-9c59-b37618033fd3\") " Dec 09 10:56:21 crc kubenswrapper[4824]: I1209 10:56:21.392316 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4d5eeca-4d1e-49d7-9c59-b37618033fd3-utilities" (OuterVolumeSpecName: "utilities") pod "a4d5eeca-4d1e-49d7-9c59-b37618033fd3" (UID: "a4d5eeca-4d1e-49d7-9c59-b37618033fd3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:56:21 crc kubenswrapper[4824]: I1209 10:56:21.403707 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4d5eeca-4d1e-49d7-9c59-b37618033fd3-kube-api-access-frzl5" (OuterVolumeSpecName: "kube-api-access-frzl5") pod "a4d5eeca-4d1e-49d7-9c59-b37618033fd3" (UID: "a4d5eeca-4d1e-49d7-9c59-b37618033fd3"). InnerVolumeSpecName "kube-api-access-frzl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 10:56:21 crc kubenswrapper[4824]: I1209 10:56:21.449203 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4d5eeca-4d1e-49d7-9c59-b37618033fd3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a4d5eeca-4d1e-49d7-9c59-b37618033fd3" (UID: "a4d5eeca-4d1e-49d7-9c59-b37618033fd3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 10:56:21 crc kubenswrapper[4824]: I1209 10:56:21.494001 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4d5eeca-4d1e-49d7-9c59-b37618033fd3-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 10:56:21 crc kubenswrapper[4824]: I1209 10:56:21.494033 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4d5eeca-4d1e-49d7-9c59-b37618033fd3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 10:56:21 crc kubenswrapper[4824]: I1209 10:56:21.494045 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frzl5\" (UniqueName: \"kubernetes.io/projected/a4d5eeca-4d1e-49d7-9c59-b37618033fd3-kube-api-access-frzl5\") on node \"crc\" DevicePath \"\"" Dec 09 10:56:22 crc kubenswrapper[4824]: I1209 10:56:22.195391 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hkjzk" Dec 09 10:56:22 crc kubenswrapper[4824]: I1209 10:56:22.245382 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hkjzk"] Dec 09 10:56:22 crc kubenswrapper[4824]: I1209 10:56:22.271386 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hkjzk"] Dec 09 10:56:23 crc kubenswrapper[4824]: I1209 10:56:23.919116 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:56:23 crc kubenswrapper[4824]: E1209 10:56:23.920304 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 10:56:23 crc kubenswrapper[4824]: I1209 10:56:23.922228 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4d5eeca-4d1e-49d7-9c59-b37618033fd3" path="/var/lib/kubelet/pods/a4d5eeca-4d1e-49d7-9c59-b37618033fd3/volumes" Dec 09 10:56:34 crc kubenswrapper[4824]: I1209 10:56:34.911582 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 10:56:35 crc kubenswrapper[4824]: I1209 10:56:35.544304 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerStarted","Data":"c5001da7621385ad85d34d158a04bc1fc9f0d5d5dcdc6b091e37b7dda74aee0b"} Dec 09 10:57:53 crc kubenswrapper[4824]: E1209 10:57:53.027996 4824 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.128:55626->38.102.83.128:33749: write tcp 38.102.83.128:55626->38.102.83.128:33749: write: broken pipe Dec 09 10:59:02 crc kubenswrapper[4824]: I1209 10:59:02.860952 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:59:02 crc kubenswrapper[4824]: I1209 10:59:02.861517 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 10:59:32 crc kubenswrapper[4824]: I1209 10:59:32.861535 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 10:59:32 crc kubenswrapper[4824]: I1209 10:59:32.862117 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:00:00 crc kubenswrapper[4824]: I1209 11:00:00.188388 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421300-rpxsw"] Dec 09 11:00:00 crc kubenswrapper[4824]: E1209 11:00:00.189693 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4d5eeca-4d1e-49d7-9c59-b37618033fd3" containerName="extract-utilities" Dec 09 11:00:00 crc kubenswrapper[4824]: I1209 11:00:00.189708 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4d5eeca-4d1e-49d7-9c59-b37618033fd3" containerName="extract-utilities" Dec 09 11:00:00 crc kubenswrapper[4824]: E1209 11:00:00.189732 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4d5eeca-4d1e-49d7-9c59-b37618033fd3" containerName="extract-content" Dec 09 11:00:00 crc kubenswrapper[4824]: I1209 11:00:00.189738 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4d5eeca-4d1e-49d7-9c59-b37618033fd3" containerName="extract-content" Dec 09 11:00:00 crc kubenswrapper[4824]: E1209 11:00:00.189762 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4d5eeca-4d1e-49d7-9c59-b37618033fd3" containerName="registry-server" Dec 09 11:00:00 crc kubenswrapper[4824]: I1209 11:00:00.189770 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4d5eeca-4d1e-49d7-9c59-b37618033fd3" containerName="registry-server" Dec 09 11:00:00 crc kubenswrapper[4824]: I1209 11:00:00.190099 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4d5eeca-4d1e-49d7-9c59-b37618033fd3" containerName="registry-server" Dec 09 11:00:00 crc kubenswrapper[4824]: I1209 11:00:00.191144 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421300-rpxsw" Dec 09 11:00:00 crc kubenswrapper[4824]: I1209 11:00:00.197373 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 11:00:00 crc kubenswrapper[4824]: I1209 11:00:00.197667 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 11:00:00 crc kubenswrapper[4824]: I1209 11:00:00.209347 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421300-rpxsw"] Dec 09 11:00:00 crc kubenswrapper[4824]: I1209 11:00:00.372422 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s77t7\" (UniqueName: \"kubernetes.io/projected/16483eab-741c-4a28-960b-2f8fdd9f0d1a-kube-api-access-s77t7\") pod \"collect-profiles-29421300-rpxsw\" (UID: \"16483eab-741c-4a28-960b-2f8fdd9f0d1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421300-rpxsw" Dec 09 11:00:00 crc kubenswrapper[4824]: I1209 11:00:00.372507 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/16483eab-741c-4a28-960b-2f8fdd9f0d1a-secret-volume\") pod \"collect-profiles-29421300-rpxsw\" (UID: \"16483eab-741c-4a28-960b-2f8fdd9f0d1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421300-rpxsw" Dec 09 11:00:00 crc kubenswrapper[4824]: I1209 11:00:00.372750 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/16483eab-741c-4a28-960b-2f8fdd9f0d1a-config-volume\") pod \"collect-profiles-29421300-rpxsw\" (UID: \"16483eab-741c-4a28-960b-2f8fdd9f0d1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421300-rpxsw" Dec 09 11:00:00 crc kubenswrapper[4824]: I1209 11:00:00.474842 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s77t7\" (UniqueName: \"kubernetes.io/projected/16483eab-741c-4a28-960b-2f8fdd9f0d1a-kube-api-access-s77t7\") pod \"collect-profiles-29421300-rpxsw\" (UID: \"16483eab-741c-4a28-960b-2f8fdd9f0d1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421300-rpxsw" Dec 09 11:00:00 crc kubenswrapper[4824]: I1209 11:00:00.474931 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/16483eab-741c-4a28-960b-2f8fdd9f0d1a-secret-volume\") pod \"collect-profiles-29421300-rpxsw\" (UID: \"16483eab-741c-4a28-960b-2f8fdd9f0d1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421300-rpxsw" Dec 09 11:00:00 crc kubenswrapper[4824]: I1209 11:00:00.475160 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/16483eab-741c-4a28-960b-2f8fdd9f0d1a-config-volume\") pod \"collect-profiles-29421300-rpxsw\" (UID: \"16483eab-741c-4a28-960b-2f8fdd9f0d1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421300-rpxsw" Dec 09 11:00:00 crc kubenswrapper[4824]: I1209 11:00:00.476321 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/16483eab-741c-4a28-960b-2f8fdd9f0d1a-config-volume\") pod \"collect-profiles-29421300-rpxsw\" (UID: \"16483eab-741c-4a28-960b-2f8fdd9f0d1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421300-rpxsw" Dec 09 11:00:00 crc kubenswrapper[4824]: I1209 11:00:00.551426 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/16483eab-741c-4a28-960b-2f8fdd9f0d1a-secret-volume\") pod \"collect-profiles-29421300-rpxsw\" (UID: \"16483eab-741c-4a28-960b-2f8fdd9f0d1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421300-rpxsw" Dec 09 11:00:00 crc kubenswrapper[4824]: I1209 11:00:00.561161 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s77t7\" (UniqueName: \"kubernetes.io/projected/16483eab-741c-4a28-960b-2f8fdd9f0d1a-kube-api-access-s77t7\") pod \"collect-profiles-29421300-rpxsw\" (UID: \"16483eab-741c-4a28-960b-2f8fdd9f0d1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421300-rpxsw" Dec 09 11:00:00 crc kubenswrapper[4824]: I1209 11:00:00.834964 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421300-rpxsw" Dec 09 11:00:01 crc kubenswrapper[4824]: I1209 11:00:01.351041 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421300-rpxsw"] Dec 09 11:00:02 crc kubenswrapper[4824]: I1209 11:00:02.082823 4824 generic.go:334] "Generic (PLEG): container finished" podID="16483eab-741c-4a28-960b-2f8fdd9f0d1a" containerID="6757e9b31af7cd49175e7a280cea4317734ce9b64b4b81b32ae5fd5824e0adf7" exitCode=0 Dec 09 11:00:02 crc kubenswrapper[4824]: I1209 11:00:02.083720 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421300-rpxsw" event={"ID":"16483eab-741c-4a28-960b-2f8fdd9f0d1a","Type":"ContainerDied","Data":"6757e9b31af7cd49175e7a280cea4317734ce9b64b4b81b32ae5fd5824e0adf7"} Dec 09 11:00:02 crc kubenswrapper[4824]: I1209 11:00:02.083839 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421300-rpxsw" event={"ID":"16483eab-741c-4a28-960b-2f8fdd9f0d1a","Type":"ContainerStarted","Data":"6977907c6e5d6ad883c70b3bbe4918fb12d9d3e0656c314162ff688b851b8806"} Dec 09 11:00:02 crc kubenswrapper[4824]: I1209 11:00:02.861504 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:00:02 crc kubenswrapper[4824]: I1209 11:00:02.861843 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:00:02 crc kubenswrapper[4824]: I1209 11:00:02.861892 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 11:00:02 crc kubenswrapper[4824]: I1209 11:00:02.864737 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c5001da7621385ad85d34d158a04bc1fc9f0d5d5dcdc6b091e37b7dda74aee0b"} pod="openshift-machine-config-operator/machine-config-daemon-dth8x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 11:00:02 crc kubenswrapper[4824]: I1209 11:00:02.864880 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" containerID="cri-o://c5001da7621385ad85d34d158a04bc1fc9f0d5d5dcdc6b091e37b7dda74aee0b" gracePeriod=600 Dec 09 11:00:03 crc kubenswrapper[4824]: I1209 11:00:03.102102 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerID="c5001da7621385ad85d34d158a04bc1fc9f0d5d5dcdc6b091e37b7dda74aee0b" exitCode=0 Dec 09 11:00:03 crc kubenswrapper[4824]: I1209 11:00:03.102157 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerDied","Data":"c5001da7621385ad85d34d158a04bc1fc9f0d5d5dcdc6b091e37b7dda74aee0b"} Dec 09 11:00:03 crc kubenswrapper[4824]: I1209 11:00:03.102465 4824 scope.go:117] "RemoveContainer" containerID="fa8a1ca5c23f205192c7d609d00671e9ba46003b82b0273d890bd305a5b8886e" Dec 09 11:00:03 crc kubenswrapper[4824]: I1209 11:00:03.632011 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421300-rpxsw" Dec 09 11:00:03 crc kubenswrapper[4824]: I1209 11:00:03.808617 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s77t7\" (UniqueName: \"kubernetes.io/projected/16483eab-741c-4a28-960b-2f8fdd9f0d1a-kube-api-access-s77t7\") pod \"16483eab-741c-4a28-960b-2f8fdd9f0d1a\" (UID: \"16483eab-741c-4a28-960b-2f8fdd9f0d1a\") " Dec 09 11:00:03 crc kubenswrapper[4824]: I1209 11:00:03.809231 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/16483eab-741c-4a28-960b-2f8fdd9f0d1a-config-volume\") pod \"16483eab-741c-4a28-960b-2f8fdd9f0d1a\" (UID: \"16483eab-741c-4a28-960b-2f8fdd9f0d1a\") " Dec 09 11:00:03 crc kubenswrapper[4824]: I1209 11:00:03.809288 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/16483eab-741c-4a28-960b-2f8fdd9f0d1a-secret-volume\") pod \"16483eab-741c-4a28-960b-2f8fdd9f0d1a\" (UID: \"16483eab-741c-4a28-960b-2f8fdd9f0d1a\") " Dec 09 11:00:03 crc kubenswrapper[4824]: I1209 11:00:03.810288 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16483eab-741c-4a28-960b-2f8fdd9f0d1a-config-volume" (OuterVolumeSpecName: "config-volume") pod "16483eab-741c-4a28-960b-2f8fdd9f0d1a" (UID: "16483eab-741c-4a28-960b-2f8fdd9f0d1a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:00:03 crc kubenswrapper[4824]: I1209 11:00:03.815746 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16483eab-741c-4a28-960b-2f8fdd9f0d1a-kube-api-access-s77t7" (OuterVolumeSpecName: "kube-api-access-s77t7") pod "16483eab-741c-4a28-960b-2f8fdd9f0d1a" (UID: "16483eab-741c-4a28-960b-2f8fdd9f0d1a"). InnerVolumeSpecName "kube-api-access-s77t7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:00:03 crc kubenswrapper[4824]: I1209 11:00:03.816603 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16483eab-741c-4a28-960b-2f8fdd9f0d1a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "16483eab-741c-4a28-960b-2f8fdd9f0d1a" (UID: "16483eab-741c-4a28-960b-2f8fdd9f0d1a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:00:03 crc kubenswrapper[4824]: I1209 11:00:03.918833 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/16483eab-741c-4a28-960b-2f8fdd9f0d1a-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 11:00:03 crc kubenswrapper[4824]: I1209 11:00:03.918873 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/16483eab-741c-4a28-960b-2f8fdd9f0d1a-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 11:00:03 crc kubenswrapper[4824]: I1209 11:00:03.918883 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s77t7\" (UniqueName: \"kubernetes.io/projected/16483eab-741c-4a28-960b-2f8fdd9f0d1a-kube-api-access-s77t7\") on node \"crc\" DevicePath \"\"" Dec 09 11:00:04 crc kubenswrapper[4824]: I1209 11:00:04.122630 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerStarted","Data":"2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4"} Dec 09 11:00:04 crc kubenswrapper[4824]: I1209 11:00:04.124974 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421300-rpxsw" event={"ID":"16483eab-741c-4a28-960b-2f8fdd9f0d1a","Type":"ContainerDied","Data":"6977907c6e5d6ad883c70b3bbe4918fb12d9d3e0656c314162ff688b851b8806"} Dec 09 11:00:04 crc kubenswrapper[4824]: I1209 11:00:04.125094 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6977907c6e5d6ad883c70b3bbe4918fb12d9d3e0656c314162ff688b851b8806" Dec 09 11:00:04 crc kubenswrapper[4824]: I1209 11:00:04.125060 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421300-rpxsw" Dec 09 11:00:04 crc kubenswrapper[4824]: I1209 11:00:04.714381 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421255-4ll88"] Dec 09 11:00:04 crc kubenswrapper[4824]: I1209 11:00:04.724975 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421255-4ll88"] Dec 09 11:00:05 crc kubenswrapper[4824]: I1209 11:00:05.924577 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1598e68-56dc-4d91-8e6f-0690ac3c285d" path="/var/lib/kubelet/pods/d1598e68-56dc-4d91-8e6f-0690ac3c285d/volumes" Dec 09 11:00:18 crc kubenswrapper[4824]: I1209 11:00:18.966428 4824 scope.go:117] "RemoveContainer" containerID="487c74959ee5ed790fab7268ff15b47a012b2c30eb1e5683d571890583b9b63c" Dec 09 11:01:00 crc kubenswrapper[4824]: I1209 11:01:00.169304 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29421301-s69hp"] Dec 09 11:01:00 crc kubenswrapper[4824]: E1209 11:01:00.171271 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16483eab-741c-4a28-960b-2f8fdd9f0d1a" containerName="collect-profiles" Dec 09 11:01:00 crc kubenswrapper[4824]: I1209 11:01:00.171308 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="16483eab-741c-4a28-960b-2f8fdd9f0d1a" containerName="collect-profiles" Dec 09 11:01:00 crc kubenswrapper[4824]: I1209 11:01:00.172004 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="16483eab-741c-4a28-960b-2f8fdd9f0d1a" containerName="collect-profiles" Dec 09 11:01:00 crc kubenswrapper[4824]: I1209 11:01:00.174089 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29421301-s69hp" Dec 09 11:01:00 crc kubenswrapper[4824]: I1209 11:01:00.184096 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29421301-s69hp"] Dec 09 11:01:00 crc kubenswrapper[4824]: I1209 11:01:00.329939 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdrz2\" (UniqueName: \"kubernetes.io/projected/818d4716-9337-4843-ac3b-9247722f5ef0-kube-api-access-xdrz2\") pod \"keystone-cron-29421301-s69hp\" (UID: \"818d4716-9337-4843-ac3b-9247722f5ef0\") " pod="openstack/keystone-cron-29421301-s69hp" Dec 09 11:01:00 crc kubenswrapper[4824]: I1209 11:01:00.330626 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/818d4716-9337-4843-ac3b-9247722f5ef0-config-data\") pod \"keystone-cron-29421301-s69hp\" (UID: \"818d4716-9337-4843-ac3b-9247722f5ef0\") " pod="openstack/keystone-cron-29421301-s69hp" Dec 09 11:01:00 crc kubenswrapper[4824]: I1209 11:01:00.330699 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/818d4716-9337-4843-ac3b-9247722f5ef0-combined-ca-bundle\") pod \"keystone-cron-29421301-s69hp\" (UID: \"818d4716-9337-4843-ac3b-9247722f5ef0\") " pod="openstack/keystone-cron-29421301-s69hp" Dec 09 11:01:00 crc kubenswrapper[4824]: I1209 11:01:00.331035 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/818d4716-9337-4843-ac3b-9247722f5ef0-fernet-keys\") pod \"keystone-cron-29421301-s69hp\" (UID: \"818d4716-9337-4843-ac3b-9247722f5ef0\") " pod="openstack/keystone-cron-29421301-s69hp" Dec 09 11:01:00 crc kubenswrapper[4824]: I1209 11:01:00.433557 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/818d4716-9337-4843-ac3b-9247722f5ef0-config-data\") pod \"keystone-cron-29421301-s69hp\" (UID: \"818d4716-9337-4843-ac3b-9247722f5ef0\") " pod="openstack/keystone-cron-29421301-s69hp" Dec 09 11:01:00 crc kubenswrapper[4824]: I1209 11:01:00.433672 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/818d4716-9337-4843-ac3b-9247722f5ef0-combined-ca-bundle\") pod \"keystone-cron-29421301-s69hp\" (UID: \"818d4716-9337-4843-ac3b-9247722f5ef0\") " pod="openstack/keystone-cron-29421301-s69hp" Dec 09 11:01:00 crc kubenswrapper[4824]: I1209 11:01:00.434047 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/818d4716-9337-4843-ac3b-9247722f5ef0-fernet-keys\") pod \"keystone-cron-29421301-s69hp\" (UID: \"818d4716-9337-4843-ac3b-9247722f5ef0\") " pod="openstack/keystone-cron-29421301-s69hp" Dec 09 11:01:00 crc kubenswrapper[4824]: I1209 11:01:00.434143 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdrz2\" (UniqueName: \"kubernetes.io/projected/818d4716-9337-4843-ac3b-9247722f5ef0-kube-api-access-xdrz2\") pod \"keystone-cron-29421301-s69hp\" (UID: \"818d4716-9337-4843-ac3b-9247722f5ef0\") " pod="openstack/keystone-cron-29421301-s69hp" Dec 09 11:01:00 crc kubenswrapper[4824]: I1209 11:01:00.440852 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/818d4716-9337-4843-ac3b-9247722f5ef0-config-data\") pod \"keystone-cron-29421301-s69hp\" (UID: \"818d4716-9337-4843-ac3b-9247722f5ef0\") " pod="openstack/keystone-cron-29421301-s69hp" Dec 09 11:01:00 crc kubenswrapper[4824]: I1209 11:01:00.442453 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/818d4716-9337-4843-ac3b-9247722f5ef0-combined-ca-bundle\") pod \"keystone-cron-29421301-s69hp\" (UID: \"818d4716-9337-4843-ac3b-9247722f5ef0\") " pod="openstack/keystone-cron-29421301-s69hp" Dec 09 11:01:00 crc kubenswrapper[4824]: I1209 11:01:00.451381 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/818d4716-9337-4843-ac3b-9247722f5ef0-fernet-keys\") pod \"keystone-cron-29421301-s69hp\" (UID: \"818d4716-9337-4843-ac3b-9247722f5ef0\") " pod="openstack/keystone-cron-29421301-s69hp" Dec 09 11:01:00 crc kubenswrapper[4824]: I1209 11:01:00.457848 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdrz2\" (UniqueName: \"kubernetes.io/projected/818d4716-9337-4843-ac3b-9247722f5ef0-kube-api-access-xdrz2\") pod \"keystone-cron-29421301-s69hp\" (UID: \"818d4716-9337-4843-ac3b-9247722f5ef0\") " pod="openstack/keystone-cron-29421301-s69hp" Dec 09 11:01:00 crc kubenswrapper[4824]: I1209 11:01:00.509462 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29421301-s69hp" Dec 09 11:01:00 crc kubenswrapper[4824]: I1209 11:01:00.968173 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29421301-s69hp"] Dec 09 11:01:01 crc kubenswrapper[4824]: I1209 11:01:01.799844 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29421301-s69hp" event={"ID":"818d4716-9337-4843-ac3b-9247722f5ef0","Type":"ContainerStarted","Data":"bede64d1fec6dd6c21bdaf9edb99dfbb0e5805b2e45b14a7e42f461399d41244"} Dec 09 11:01:01 crc kubenswrapper[4824]: I1209 11:01:01.801094 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29421301-s69hp" event={"ID":"818d4716-9337-4843-ac3b-9247722f5ef0","Type":"ContainerStarted","Data":"ca831fa1cf84f5fff81edc9c43e8107da35a6126652e98cf0593d48b486e7314"} Dec 09 11:01:01 crc kubenswrapper[4824]: I1209 11:01:01.833456 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29421301-s69hp" podStartSLOduration=1.833435421 podStartE2EDuration="1.833435421s" podCreationTimestamp="2025-12-09 11:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 11:01:01.830366005 +0000 UTC m=+4418.164870672" watchObservedRunningTime="2025-12-09 11:01:01.833435421 +0000 UTC m=+4418.167940088" Dec 09 11:01:03 crc kubenswrapper[4824]: I1209 11:01:03.647866 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-227pt"] Dec 09 11:01:03 crc kubenswrapper[4824]: I1209 11:01:03.651414 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-227pt" Dec 09 11:01:03 crc kubenswrapper[4824]: I1209 11:01:03.662410 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-227pt"] Dec 09 11:01:03 crc kubenswrapper[4824]: I1209 11:01:03.734394 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcb2068f-bdd4-497b-99a6-5c46099c240a-utilities\") pod \"redhat-marketplace-227pt\" (UID: \"fcb2068f-bdd4-497b-99a6-5c46099c240a\") " pod="openshift-marketplace/redhat-marketplace-227pt" Dec 09 11:01:03 crc kubenswrapper[4824]: I1209 11:01:03.734619 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcb2068f-bdd4-497b-99a6-5c46099c240a-catalog-content\") pod \"redhat-marketplace-227pt\" (UID: \"fcb2068f-bdd4-497b-99a6-5c46099c240a\") " pod="openshift-marketplace/redhat-marketplace-227pt" Dec 09 11:01:03 crc kubenswrapper[4824]: I1209 11:01:03.734807 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldcst\" (UniqueName: \"kubernetes.io/projected/fcb2068f-bdd4-497b-99a6-5c46099c240a-kube-api-access-ldcst\") pod \"redhat-marketplace-227pt\" (UID: \"fcb2068f-bdd4-497b-99a6-5c46099c240a\") " pod="openshift-marketplace/redhat-marketplace-227pt" Dec 09 11:01:03 crc kubenswrapper[4824]: I1209 11:01:03.836968 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcb2068f-bdd4-497b-99a6-5c46099c240a-utilities\") pod \"redhat-marketplace-227pt\" (UID: \"fcb2068f-bdd4-497b-99a6-5c46099c240a\") " pod="openshift-marketplace/redhat-marketplace-227pt" Dec 09 11:01:03 crc kubenswrapper[4824]: I1209 11:01:03.837363 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcb2068f-bdd4-497b-99a6-5c46099c240a-catalog-content\") pod \"redhat-marketplace-227pt\" (UID: \"fcb2068f-bdd4-497b-99a6-5c46099c240a\") " pod="openshift-marketplace/redhat-marketplace-227pt" Dec 09 11:01:03 crc kubenswrapper[4824]: I1209 11:01:03.837523 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldcst\" (UniqueName: \"kubernetes.io/projected/fcb2068f-bdd4-497b-99a6-5c46099c240a-kube-api-access-ldcst\") pod \"redhat-marketplace-227pt\" (UID: \"fcb2068f-bdd4-497b-99a6-5c46099c240a\") " pod="openshift-marketplace/redhat-marketplace-227pt" Dec 09 11:01:03 crc kubenswrapper[4824]: I1209 11:01:03.837521 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcb2068f-bdd4-497b-99a6-5c46099c240a-utilities\") pod \"redhat-marketplace-227pt\" (UID: \"fcb2068f-bdd4-497b-99a6-5c46099c240a\") " pod="openshift-marketplace/redhat-marketplace-227pt" Dec 09 11:01:03 crc kubenswrapper[4824]: I1209 11:01:03.837873 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcb2068f-bdd4-497b-99a6-5c46099c240a-catalog-content\") pod \"redhat-marketplace-227pt\" (UID: \"fcb2068f-bdd4-497b-99a6-5c46099c240a\") " pod="openshift-marketplace/redhat-marketplace-227pt" Dec 09 11:01:03 crc kubenswrapper[4824]: I1209 11:01:03.871712 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldcst\" (UniqueName: \"kubernetes.io/projected/fcb2068f-bdd4-497b-99a6-5c46099c240a-kube-api-access-ldcst\") pod \"redhat-marketplace-227pt\" (UID: \"fcb2068f-bdd4-497b-99a6-5c46099c240a\") " pod="openshift-marketplace/redhat-marketplace-227pt" Dec 09 11:01:03 crc kubenswrapper[4824]: I1209 11:01:03.981257 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-227pt" Dec 09 11:01:05 crc kubenswrapper[4824]: I1209 11:01:05.535671 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-227pt"] Dec 09 11:01:05 crc kubenswrapper[4824]: W1209 11:01:05.539109 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfcb2068f_bdd4_497b_99a6_5c46099c240a.slice/crio-1b1568f524b7ec6ea887f2e9516522eb7dc66dfe8676c576d69cb9e9cc6e9ab2 WatchSource:0}: Error finding container 1b1568f524b7ec6ea887f2e9516522eb7dc66dfe8676c576d69cb9e9cc6e9ab2: Status 404 returned error can't find the container with id 1b1568f524b7ec6ea887f2e9516522eb7dc66dfe8676c576d69cb9e9cc6e9ab2 Dec 09 11:01:05 crc kubenswrapper[4824]: I1209 11:01:05.867544 4824 generic.go:334] "Generic (PLEG): container finished" podID="818d4716-9337-4843-ac3b-9247722f5ef0" containerID="bede64d1fec6dd6c21bdaf9edb99dfbb0e5805b2e45b14a7e42f461399d41244" exitCode=0 Dec 09 11:01:05 crc kubenswrapper[4824]: I1209 11:01:05.867604 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29421301-s69hp" event={"ID":"818d4716-9337-4843-ac3b-9247722f5ef0","Type":"ContainerDied","Data":"bede64d1fec6dd6c21bdaf9edb99dfbb0e5805b2e45b14a7e42f461399d41244"} Dec 09 11:01:05 crc kubenswrapper[4824]: I1209 11:01:05.871872 4824 generic.go:334] "Generic (PLEG): container finished" podID="fcb2068f-bdd4-497b-99a6-5c46099c240a" containerID="021836bbbe87ae8318ef8bc736921e91f4b65c557e74baeceaf22e2f74a207f6" exitCode=0 Dec 09 11:01:05 crc kubenswrapper[4824]: I1209 11:01:05.872252 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-227pt" event={"ID":"fcb2068f-bdd4-497b-99a6-5c46099c240a","Type":"ContainerDied","Data":"021836bbbe87ae8318ef8bc736921e91f4b65c557e74baeceaf22e2f74a207f6"} Dec 09 11:01:05 crc kubenswrapper[4824]: I1209 11:01:05.872330 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-227pt" event={"ID":"fcb2068f-bdd4-497b-99a6-5c46099c240a","Type":"ContainerStarted","Data":"1b1568f524b7ec6ea887f2e9516522eb7dc66dfe8676c576d69cb9e9cc6e9ab2"} Dec 09 11:01:06 crc kubenswrapper[4824]: I1209 11:01:06.885107 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-227pt" event={"ID":"fcb2068f-bdd4-497b-99a6-5c46099c240a","Type":"ContainerStarted","Data":"3bca54407a1f689f91448cd3fa702554ba296d21d9141aae77ee5a70b0de5d17"} Dec 09 11:01:07 crc kubenswrapper[4824]: I1209 11:01:07.313734 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29421301-s69hp" Dec 09 11:01:07 crc kubenswrapper[4824]: I1209 11:01:07.436113 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/818d4716-9337-4843-ac3b-9247722f5ef0-config-data\") pod \"818d4716-9337-4843-ac3b-9247722f5ef0\" (UID: \"818d4716-9337-4843-ac3b-9247722f5ef0\") " Dec 09 11:01:07 crc kubenswrapper[4824]: I1209 11:01:07.436342 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdrz2\" (UniqueName: \"kubernetes.io/projected/818d4716-9337-4843-ac3b-9247722f5ef0-kube-api-access-xdrz2\") pod \"818d4716-9337-4843-ac3b-9247722f5ef0\" (UID: \"818d4716-9337-4843-ac3b-9247722f5ef0\") " Dec 09 11:01:07 crc kubenswrapper[4824]: I1209 11:01:07.436478 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/818d4716-9337-4843-ac3b-9247722f5ef0-combined-ca-bundle\") pod \"818d4716-9337-4843-ac3b-9247722f5ef0\" (UID: \"818d4716-9337-4843-ac3b-9247722f5ef0\") " Dec 09 11:01:07 crc kubenswrapper[4824]: I1209 11:01:07.436556 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/818d4716-9337-4843-ac3b-9247722f5ef0-fernet-keys\") pod \"818d4716-9337-4843-ac3b-9247722f5ef0\" (UID: \"818d4716-9337-4843-ac3b-9247722f5ef0\") " Dec 09 11:01:07 crc kubenswrapper[4824]: I1209 11:01:07.442497 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/818d4716-9337-4843-ac3b-9247722f5ef0-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "818d4716-9337-4843-ac3b-9247722f5ef0" (UID: "818d4716-9337-4843-ac3b-9247722f5ef0"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:01:07 crc kubenswrapper[4824]: I1209 11:01:07.450977 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/818d4716-9337-4843-ac3b-9247722f5ef0-kube-api-access-xdrz2" (OuterVolumeSpecName: "kube-api-access-xdrz2") pod "818d4716-9337-4843-ac3b-9247722f5ef0" (UID: "818d4716-9337-4843-ac3b-9247722f5ef0"). InnerVolumeSpecName "kube-api-access-xdrz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:01:07 crc kubenswrapper[4824]: I1209 11:01:07.472694 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/818d4716-9337-4843-ac3b-9247722f5ef0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "818d4716-9337-4843-ac3b-9247722f5ef0" (UID: "818d4716-9337-4843-ac3b-9247722f5ef0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:01:07 crc kubenswrapper[4824]: I1209 11:01:07.513505 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/818d4716-9337-4843-ac3b-9247722f5ef0-config-data" (OuterVolumeSpecName: "config-data") pod "818d4716-9337-4843-ac3b-9247722f5ef0" (UID: "818d4716-9337-4843-ac3b-9247722f5ef0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:01:07 crc kubenswrapper[4824]: I1209 11:01:07.540068 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/818d4716-9337-4843-ac3b-9247722f5ef0-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:01:07 crc kubenswrapper[4824]: I1209 11:01:07.540108 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdrz2\" (UniqueName: \"kubernetes.io/projected/818d4716-9337-4843-ac3b-9247722f5ef0-kube-api-access-xdrz2\") on node \"crc\" DevicePath \"\"" Dec 09 11:01:07 crc kubenswrapper[4824]: I1209 11:01:07.540119 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/818d4716-9337-4843-ac3b-9247722f5ef0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 11:01:07 crc kubenswrapper[4824]: I1209 11:01:07.540130 4824 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/818d4716-9337-4843-ac3b-9247722f5ef0-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 09 11:01:07 crc kubenswrapper[4824]: I1209 11:01:07.898486 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29421301-s69hp" Dec 09 11:01:07 crc kubenswrapper[4824]: I1209 11:01:07.898525 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29421301-s69hp" event={"ID":"818d4716-9337-4843-ac3b-9247722f5ef0","Type":"ContainerDied","Data":"ca831fa1cf84f5fff81edc9c43e8107da35a6126652e98cf0593d48b486e7314"} Dec 09 11:01:07 crc kubenswrapper[4824]: I1209 11:01:07.899670 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca831fa1cf84f5fff81edc9c43e8107da35a6126652e98cf0593d48b486e7314" Dec 09 11:01:07 crc kubenswrapper[4824]: I1209 11:01:07.901289 4824 generic.go:334] "Generic (PLEG): container finished" podID="fcb2068f-bdd4-497b-99a6-5c46099c240a" containerID="3bca54407a1f689f91448cd3fa702554ba296d21d9141aae77ee5a70b0de5d17" exitCode=0 Dec 09 11:01:07 crc kubenswrapper[4824]: I1209 11:01:07.901341 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-227pt" event={"ID":"fcb2068f-bdd4-497b-99a6-5c46099c240a","Type":"ContainerDied","Data":"3bca54407a1f689f91448cd3fa702554ba296d21d9141aae77ee5a70b0de5d17"} Dec 09 11:01:09 crc kubenswrapper[4824]: I1209 11:01:09.933404 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-227pt" event={"ID":"fcb2068f-bdd4-497b-99a6-5c46099c240a","Type":"ContainerStarted","Data":"895037bfcc15e47623201ab1830e5074064c46e0c3ddc915dc716ea4d45f1e2f"} Dec 09 11:01:09 crc kubenswrapper[4824]: I1209 11:01:09.968031 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-227pt" podStartSLOduration=4.022376486 podStartE2EDuration="6.968005203s" podCreationTimestamp="2025-12-09 11:01:03 +0000 UTC" firstStartedPulling="2025-12-09 11:01:05.874290906 +0000 UTC m=+4422.208795573" lastFinishedPulling="2025-12-09 11:01:08.819919623 +0000 UTC m=+4425.154424290" observedRunningTime="2025-12-09 11:01:09.953663373 +0000 UTC m=+4426.288168040" watchObservedRunningTime="2025-12-09 11:01:09.968005203 +0000 UTC m=+4426.302509880" Dec 09 11:01:13 crc kubenswrapper[4824]: I1209 11:01:13.982080 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-227pt" Dec 09 11:01:13 crc kubenswrapper[4824]: I1209 11:01:13.983918 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-227pt" Dec 09 11:01:14 crc kubenswrapper[4824]: I1209 11:01:14.044420 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-227pt" Dec 09 11:01:15 crc kubenswrapper[4824]: I1209 11:01:15.038317 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-227pt" Dec 09 11:01:15 crc kubenswrapper[4824]: I1209 11:01:15.094530 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-227pt"] Dec 09 11:01:17 crc kubenswrapper[4824]: I1209 11:01:17.017359 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-227pt" podUID="fcb2068f-bdd4-497b-99a6-5c46099c240a" containerName="registry-server" containerID="cri-o://895037bfcc15e47623201ab1830e5074064c46e0c3ddc915dc716ea4d45f1e2f" gracePeriod=2 Dec 09 11:01:17 crc kubenswrapper[4824]: I1209 11:01:17.798959 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-227pt" Dec 09 11:01:17 crc kubenswrapper[4824]: I1209 11:01:17.962511 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcb2068f-bdd4-497b-99a6-5c46099c240a-utilities\") pod \"fcb2068f-bdd4-497b-99a6-5c46099c240a\" (UID: \"fcb2068f-bdd4-497b-99a6-5c46099c240a\") " Dec 09 11:01:17 crc kubenswrapper[4824]: I1209 11:01:17.962775 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldcst\" (UniqueName: \"kubernetes.io/projected/fcb2068f-bdd4-497b-99a6-5c46099c240a-kube-api-access-ldcst\") pod \"fcb2068f-bdd4-497b-99a6-5c46099c240a\" (UID: \"fcb2068f-bdd4-497b-99a6-5c46099c240a\") " Dec 09 11:01:17 crc kubenswrapper[4824]: I1209 11:01:17.962858 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcb2068f-bdd4-497b-99a6-5c46099c240a-catalog-content\") pod \"fcb2068f-bdd4-497b-99a6-5c46099c240a\" (UID: \"fcb2068f-bdd4-497b-99a6-5c46099c240a\") " Dec 09 11:01:17 crc kubenswrapper[4824]: I1209 11:01:17.963725 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcb2068f-bdd4-497b-99a6-5c46099c240a-utilities" (OuterVolumeSpecName: "utilities") pod "fcb2068f-bdd4-497b-99a6-5c46099c240a" (UID: "fcb2068f-bdd4-497b-99a6-5c46099c240a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:01:17 crc kubenswrapper[4824]: I1209 11:01:17.971128 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcb2068f-bdd4-497b-99a6-5c46099c240a-kube-api-access-ldcst" (OuterVolumeSpecName: "kube-api-access-ldcst") pod "fcb2068f-bdd4-497b-99a6-5c46099c240a" (UID: "fcb2068f-bdd4-497b-99a6-5c46099c240a"). InnerVolumeSpecName "kube-api-access-ldcst". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:01:17 crc kubenswrapper[4824]: I1209 11:01:17.990535 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcb2068f-bdd4-497b-99a6-5c46099c240a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fcb2068f-bdd4-497b-99a6-5c46099c240a" (UID: "fcb2068f-bdd4-497b-99a6-5c46099c240a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:01:18 crc kubenswrapper[4824]: I1209 11:01:18.030461 4824 generic.go:334] "Generic (PLEG): container finished" podID="fcb2068f-bdd4-497b-99a6-5c46099c240a" containerID="895037bfcc15e47623201ab1830e5074064c46e0c3ddc915dc716ea4d45f1e2f" exitCode=0 Dec 09 11:01:18 crc kubenswrapper[4824]: I1209 11:01:18.030501 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-227pt" event={"ID":"fcb2068f-bdd4-497b-99a6-5c46099c240a","Type":"ContainerDied","Data":"895037bfcc15e47623201ab1830e5074064c46e0c3ddc915dc716ea4d45f1e2f"} Dec 09 11:01:18 crc kubenswrapper[4824]: I1209 11:01:18.030526 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-227pt" event={"ID":"fcb2068f-bdd4-497b-99a6-5c46099c240a","Type":"ContainerDied","Data":"1b1568f524b7ec6ea887f2e9516522eb7dc66dfe8676c576d69cb9e9cc6e9ab2"} Dec 09 11:01:18 crc kubenswrapper[4824]: I1209 11:01:18.030544 4824 scope.go:117] "RemoveContainer" containerID="895037bfcc15e47623201ab1830e5074064c46e0c3ddc915dc716ea4d45f1e2f" Dec 09 11:01:18 crc kubenswrapper[4824]: I1209 11:01:18.030673 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-227pt" Dec 09 11:01:18 crc kubenswrapper[4824]: I1209 11:01:18.072497 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcb2068f-bdd4-497b-99a6-5c46099c240a-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:01:18 crc kubenswrapper[4824]: I1209 11:01:18.072535 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldcst\" (UniqueName: \"kubernetes.io/projected/fcb2068f-bdd4-497b-99a6-5c46099c240a-kube-api-access-ldcst\") on node \"crc\" DevicePath \"\"" Dec 09 11:01:18 crc kubenswrapper[4824]: I1209 11:01:18.072545 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcb2068f-bdd4-497b-99a6-5c46099c240a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:01:18 crc kubenswrapper[4824]: I1209 11:01:18.075404 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-227pt"] Dec 09 11:01:18 crc kubenswrapper[4824]: I1209 11:01:18.080008 4824 scope.go:117] "RemoveContainer" containerID="3bca54407a1f689f91448cd3fa702554ba296d21d9141aae77ee5a70b0de5d17" Dec 09 11:01:18 crc kubenswrapper[4824]: I1209 11:01:18.093900 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-227pt"] Dec 09 11:01:18 crc kubenswrapper[4824]: I1209 11:01:18.109459 4824 scope.go:117] "RemoveContainer" containerID="021836bbbe87ae8318ef8bc736921e91f4b65c557e74baeceaf22e2f74a207f6" Dec 09 11:01:18 crc kubenswrapper[4824]: I1209 11:01:18.183759 4824 scope.go:117] "RemoveContainer" containerID="895037bfcc15e47623201ab1830e5074064c46e0c3ddc915dc716ea4d45f1e2f" Dec 09 11:01:18 crc kubenswrapper[4824]: E1209 11:01:18.184367 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"895037bfcc15e47623201ab1830e5074064c46e0c3ddc915dc716ea4d45f1e2f\": container with ID starting with 895037bfcc15e47623201ab1830e5074064c46e0c3ddc915dc716ea4d45f1e2f not found: ID does not exist" containerID="895037bfcc15e47623201ab1830e5074064c46e0c3ddc915dc716ea4d45f1e2f" Dec 09 11:01:18 crc kubenswrapper[4824]: I1209 11:01:18.184436 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"895037bfcc15e47623201ab1830e5074064c46e0c3ddc915dc716ea4d45f1e2f"} err="failed to get container status \"895037bfcc15e47623201ab1830e5074064c46e0c3ddc915dc716ea4d45f1e2f\": rpc error: code = NotFound desc = could not find container \"895037bfcc15e47623201ab1830e5074064c46e0c3ddc915dc716ea4d45f1e2f\": container with ID starting with 895037bfcc15e47623201ab1830e5074064c46e0c3ddc915dc716ea4d45f1e2f not found: ID does not exist" Dec 09 11:01:18 crc kubenswrapper[4824]: I1209 11:01:18.184474 4824 scope.go:117] "RemoveContainer" containerID="3bca54407a1f689f91448cd3fa702554ba296d21d9141aae77ee5a70b0de5d17" Dec 09 11:01:18 crc kubenswrapper[4824]: E1209 11:01:18.185015 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bca54407a1f689f91448cd3fa702554ba296d21d9141aae77ee5a70b0de5d17\": container with ID starting with 3bca54407a1f689f91448cd3fa702554ba296d21d9141aae77ee5a70b0de5d17 not found: ID does not exist" containerID="3bca54407a1f689f91448cd3fa702554ba296d21d9141aae77ee5a70b0de5d17" Dec 09 11:01:18 crc kubenswrapper[4824]: I1209 11:01:18.185069 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bca54407a1f689f91448cd3fa702554ba296d21d9141aae77ee5a70b0de5d17"} err="failed to get container status \"3bca54407a1f689f91448cd3fa702554ba296d21d9141aae77ee5a70b0de5d17\": rpc error: code = NotFound desc = could not find container \"3bca54407a1f689f91448cd3fa702554ba296d21d9141aae77ee5a70b0de5d17\": container with ID starting with 3bca54407a1f689f91448cd3fa702554ba296d21d9141aae77ee5a70b0de5d17 not found: ID does not exist" Dec 09 11:01:18 crc kubenswrapper[4824]: I1209 11:01:18.185096 4824 scope.go:117] "RemoveContainer" containerID="021836bbbe87ae8318ef8bc736921e91f4b65c557e74baeceaf22e2f74a207f6" Dec 09 11:01:18 crc kubenswrapper[4824]: E1209 11:01:18.185919 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"021836bbbe87ae8318ef8bc736921e91f4b65c557e74baeceaf22e2f74a207f6\": container with ID starting with 021836bbbe87ae8318ef8bc736921e91f4b65c557e74baeceaf22e2f74a207f6 not found: ID does not exist" containerID="021836bbbe87ae8318ef8bc736921e91f4b65c557e74baeceaf22e2f74a207f6" Dec 09 11:01:18 crc kubenswrapper[4824]: I1209 11:01:18.185952 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"021836bbbe87ae8318ef8bc736921e91f4b65c557e74baeceaf22e2f74a207f6"} err="failed to get container status \"021836bbbe87ae8318ef8bc736921e91f4b65c557e74baeceaf22e2f74a207f6\": rpc error: code = NotFound desc = could not find container \"021836bbbe87ae8318ef8bc736921e91f4b65c557e74baeceaf22e2f74a207f6\": container with ID starting with 021836bbbe87ae8318ef8bc736921e91f4b65c557e74baeceaf22e2f74a207f6 not found: ID does not exist" Dec 09 11:01:19 crc kubenswrapper[4824]: I1209 11:01:19.924721 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcb2068f-bdd4-497b-99a6-5c46099c240a" path="/var/lib/kubelet/pods/fcb2068f-bdd4-497b-99a6-5c46099c240a/volumes" Dec 09 11:02:09 crc kubenswrapper[4824]: I1209 11:02:09.637039 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5wvbr"] Dec 09 11:02:09 crc kubenswrapper[4824]: E1209 11:02:09.638361 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcb2068f-bdd4-497b-99a6-5c46099c240a" containerName="registry-server" Dec 09 11:02:09 crc kubenswrapper[4824]: I1209 11:02:09.638381 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcb2068f-bdd4-497b-99a6-5c46099c240a" containerName="registry-server" Dec 09 11:02:09 crc kubenswrapper[4824]: E1209 11:02:09.638456 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcb2068f-bdd4-497b-99a6-5c46099c240a" containerName="extract-utilities" Dec 09 11:02:09 crc kubenswrapper[4824]: I1209 11:02:09.638466 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcb2068f-bdd4-497b-99a6-5c46099c240a" containerName="extract-utilities" Dec 09 11:02:09 crc kubenswrapper[4824]: E1209 11:02:09.638487 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="818d4716-9337-4843-ac3b-9247722f5ef0" containerName="keystone-cron" Dec 09 11:02:09 crc kubenswrapper[4824]: I1209 11:02:09.638496 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="818d4716-9337-4843-ac3b-9247722f5ef0" containerName="keystone-cron" Dec 09 11:02:09 crc kubenswrapper[4824]: E1209 11:02:09.638514 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcb2068f-bdd4-497b-99a6-5c46099c240a" containerName="extract-content" Dec 09 11:02:09 crc kubenswrapper[4824]: I1209 11:02:09.638521 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcb2068f-bdd4-497b-99a6-5c46099c240a" containerName="extract-content" Dec 09 11:02:09 crc kubenswrapper[4824]: I1209 11:02:09.638931 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="818d4716-9337-4843-ac3b-9247722f5ef0" containerName="keystone-cron" Dec 09 11:02:09 crc kubenswrapper[4824]: I1209 11:02:09.638958 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcb2068f-bdd4-497b-99a6-5c46099c240a" containerName="registry-server" Dec 09 11:02:09 crc kubenswrapper[4824]: I1209 11:02:09.641925 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5wvbr" Dec 09 11:02:09 crc kubenswrapper[4824]: I1209 11:02:09.653334 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5wvbr"] Dec 09 11:02:09 crc kubenswrapper[4824]: I1209 11:02:09.829209 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b350ac3d-9036-421f-aade-102c1737e38e-catalog-content\") pod \"certified-operators-5wvbr\" (UID: \"b350ac3d-9036-421f-aade-102c1737e38e\") " pod="openshift-marketplace/certified-operators-5wvbr" Dec 09 11:02:09 crc kubenswrapper[4824]: I1209 11:02:09.829493 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b350ac3d-9036-421f-aade-102c1737e38e-utilities\") pod \"certified-operators-5wvbr\" (UID: \"b350ac3d-9036-421f-aade-102c1737e38e\") " pod="openshift-marketplace/certified-operators-5wvbr" Dec 09 11:02:09 crc kubenswrapper[4824]: I1209 11:02:09.829577 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skgdf\" (UniqueName: \"kubernetes.io/projected/b350ac3d-9036-421f-aade-102c1737e38e-kube-api-access-skgdf\") pod \"certified-operators-5wvbr\" (UID: \"b350ac3d-9036-421f-aade-102c1737e38e\") " pod="openshift-marketplace/certified-operators-5wvbr" Dec 09 11:02:09 crc kubenswrapper[4824]: I1209 11:02:09.932949 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b350ac3d-9036-421f-aade-102c1737e38e-catalog-content\") pod \"certified-operators-5wvbr\" (UID: \"b350ac3d-9036-421f-aade-102c1737e38e\") " pod="openshift-marketplace/certified-operators-5wvbr" Dec 09 11:02:09 crc kubenswrapper[4824]: I1209 11:02:09.933044 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b350ac3d-9036-421f-aade-102c1737e38e-utilities\") pod \"certified-operators-5wvbr\" (UID: \"b350ac3d-9036-421f-aade-102c1737e38e\") " pod="openshift-marketplace/certified-operators-5wvbr" Dec 09 11:02:09 crc kubenswrapper[4824]: I1209 11:02:09.933075 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skgdf\" (UniqueName: \"kubernetes.io/projected/b350ac3d-9036-421f-aade-102c1737e38e-kube-api-access-skgdf\") pod \"certified-operators-5wvbr\" (UID: \"b350ac3d-9036-421f-aade-102c1737e38e\") " pod="openshift-marketplace/certified-operators-5wvbr" Dec 09 11:02:09 crc kubenswrapper[4824]: I1209 11:02:09.933377 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b350ac3d-9036-421f-aade-102c1737e38e-catalog-content\") pod \"certified-operators-5wvbr\" (UID: \"b350ac3d-9036-421f-aade-102c1737e38e\") " pod="openshift-marketplace/certified-operators-5wvbr" Dec 09 11:02:09 crc kubenswrapper[4824]: I1209 11:02:09.933600 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b350ac3d-9036-421f-aade-102c1737e38e-utilities\") pod \"certified-operators-5wvbr\" (UID: \"b350ac3d-9036-421f-aade-102c1737e38e\") " pod="openshift-marketplace/certified-operators-5wvbr" Dec 09 11:02:09 crc kubenswrapper[4824]: I1209 11:02:09.954047 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skgdf\" (UniqueName: \"kubernetes.io/projected/b350ac3d-9036-421f-aade-102c1737e38e-kube-api-access-skgdf\") pod \"certified-operators-5wvbr\" (UID: \"b350ac3d-9036-421f-aade-102c1737e38e\") " pod="openshift-marketplace/certified-operators-5wvbr" Dec 09 11:02:09 crc kubenswrapper[4824]: I1209 11:02:09.978303 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5wvbr" Dec 09 11:02:10 crc kubenswrapper[4824]: I1209 11:02:10.581405 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5wvbr"] Dec 09 11:02:10 crc kubenswrapper[4824]: I1209 11:02:10.718555 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5wvbr" event={"ID":"b350ac3d-9036-421f-aade-102c1737e38e","Type":"ContainerStarted","Data":"fb03ce3a34e295a2e7f2be9e451cf31899108b9a7b41900e8d90e1c11ac7b5d4"} Dec 09 11:02:11 crc kubenswrapper[4824]: I1209 11:02:11.738880 4824 generic.go:334] "Generic (PLEG): container finished" podID="b350ac3d-9036-421f-aade-102c1737e38e" containerID="b7a2c468892d2592e913149bc6bb4c054ca501c492d14d81213a818e78c12e87" exitCode=0 Dec 09 11:02:11 crc kubenswrapper[4824]: I1209 11:02:11.739372 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5wvbr" event={"ID":"b350ac3d-9036-421f-aade-102c1737e38e","Type":"ContainerDied","Data":"b7a2c468892d2592e913149bc6bb4c054ca501c492d14d81213a818e78c12e87"} Dec 09 11:02:11 crc kubenswrapper[4824]: I1209 11:02:11.743219 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 11:02:13 crc kubenswrapper[4824]: I1209 11:02:13.768809 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5wvbr" event={"ID":"b350ac3d-9036-421f-aade-102c1737e38e","Type":"ContainerStarted","Data":"3a9e3735e7d17f718166b9d859c00bb7ced6f5da1be983096fa9e8ce3565db66"} Dec 09 11:02:14 crc kubenswrapper[4824]: I1209 11:02:14.782858 4824 generic.go:334] "Generic (PLEG): container finished" podID="b350ac3d-9036-421f-aade-102c1737e38e" containerID="3a9e3735e7d17f718166b9d859c00bb7ced6f5da1be983096fa9e8ce3565db66" exitCode=0 Dec 09 11:02:14 crc kubenswrapper[4824]: I1209 11:02:14.782923 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5wvbr" event={"ID":"b350ac3d-9036-421f-aade-102c1737e38e","Type":"ContainerDied","Data":"3a9e3735e7d17f718166b9d859c00bb7ced6f5da1be983096fa9e8ce3565db66"} Dec 09 11:02:15 crc kubenswrapper[4824]: I1209 11:02:15.799003 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5wvbr" event={"ID":"b350ac3d-9036-421f-aade-102c1737e38e","Type":"ContainerStarted","Data":"fc717ac53c7847e8ad1a1e98fc5066a0839ada0187706933e40d4b754c873d22"} Dec 09 11:02:19 crc kubenswrapper[4824]: I1209 11:02:19.184194 4824 scope.go:117] "RemoveContainer" containerID="686f152adecbc4a01864dd6e995dd9a27862affc114fb21409f443642d4c9c5c" Dec 09 11:02:19 crc kubenswrapper[4824]: I1209 11:02:19.250117 4824 scope.go:117] "RemoveContainer" containerID="20e34433aa2f535e862c1e9c7d942ae34d44e43491d3029734fbe1aab5c7312d" Dec 09 11:02:19 crc kubenswrapper[4824]: I1209 11:02:19.354940 4824 scope.go:117] "RemoveContainer" containerID="ad894e1b364c9e3f8c9a0c6037a4481953769d4414bb08453aead36c1c040e7e" Dec 09 11:02:19 crc kubenswrapper[4824]: I1209 11:02:19.978951 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5wvbr" Dec 09 11:02:19 crc kubenswrapper[4824]: I1209 11:02:19.979018 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5wvbr" Dec 09 11:02:20 crc kubenswrapper[4824]: I1209 11:02:20.035567 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5wvbr" Dec 09 11:02:20 crc kubenswrapper[4824]: I1209 11:02:20.054482 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5wvbr" podStartSLOduration=7.600259166 podStartE2EDuration="11.05446201s" podCreationTimestamp="2025-12-09 11:02:09 +0000 UTC" firstStartedPulling="2025-12-09 11:02:11.742916057 +0000 UTC m=+4488.077420724" lastFinishedPulling="2025-12-09 11:02:15.197118901 +0000 UTC m=+4491.531623568" observedRunningTime="2025-12-09 11:02:15.822234151 +0000 UTC m=+4492.156738838" watchObservedRunningTime="2025-12-09 11:02:20.05446201 +0000 UTC m=+4496.388966667" Dec 09 11:02:21 crc kubenswrapper[4824]: I1209 11:02:21.213378 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5wvbr" Dec 09 11:02:21 crc kubenswrapper[4824]: I1209 11:02:21.286050 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5wvbr"] Dec 09 11:02:22 crc kubenswrapper[4824]: I1209 11:02:22.875602 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5wvbr" podUID="b350ac3d-9036-421f-aade-102c1737e38e" containerName="registry-server" containerID="cri-o://fc717ac53c7847e8ad1a1e98fc5066a0839ada0187706933e40d4b754c873d22" gracePeriod=2 Dec 09 11:02:23 crc kubenswrapper[4824]: I1209 11:02:23.409377 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5wvbr" Dec 09 11:02:23 crc kubenswrapper[4824]: I1209 11:02:23.494379 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skgdf\" (UniqueName: \"kubernetes.io/projected/b350ac3d-9036-421f-aade-102c1737e38e-kube-api-access-skgdf\") pod \"b350ac3d-9036-421f-aade-102c1737e38e\" (UID: \"b350ac3d-9036-421f-aade-102c1737e38e\") " Dec 09 11:02:23 crc kubenswrapper[4824]: I1209 11:02:23.494459 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b350ac3d-9036-421f-aade-102c1737e38e-catalog-content\") pod \"b350ac3d-9036-421f-aade-102c1737e38e\" (UID: \"b350ac3d-9036-421f-aade-102c1737e38e\") " Dec 09 11:02:23 crc kubenswrapper[4824]: I1209 11:02:23.494721 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b350ac3d-9036-421f-aade-102c1737e38e-utilities\") pod \"b350ac3d-9036-421f-aade-102c1737e38e\" (UID: \"b350ac3d-9036-421f-aade-102c1737e38e\") " Dec 09 11:02:23 crc kubenswrapper[4824]: I1209 11:02:23.495577 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b350ac3d-9036-421f-aade-102c1737e38e-utilities" (OuterVolumeSpecName: "utilities") pod "b350ac3d-9036-421f-aade-102c1737e38e" (UID: "b350ac3d-9036-421f-aade-102c1737e38e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:02:23 crc kubenswrapper[4824]: I1209 11:02:23.515352 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b350ac3d-9036-421f-aade-102c1737e38e-kube-api-access-skgdf" (OuterVolumeSpecName: "kube-api-access-skgdf") pod "b350ac3d-9036-421f-aade-102c1737e38e" (UID: "b350ac3d-9036-421f-aade-102c1737e38e"). InnerVolumeSpecName "kube-api-access-skgdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:02:23 crc kubenswrapper[4824]: I1209 11:02:23.583506 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b350ac3d-9036-421f-aade-102c1737e38e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b350ac3d-9036-421f-aade-102c1737e38e" (UID: "b350ac3d-9036-421f-aade-102c1737e38e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:02:23 crc kubenswrapper[4824]: I1209 11:02:23.598085 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skgdf\" (UniqueName: \"kubernetes.io/projected/b350ac3d-9036-421f-aade-102c1737e38e-kube-api-access-skgdf\") on node \"crc\" DevicePath \"\"" Dec 09 11:02:23 crc kubenswrapper[4824]: I1209 11:02:23.598135 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b350ac3d-9036-421f-aade-102c1737e38e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:02:23 crc kubenswrapper[4824]: I1209 11:02:23.598151 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b350ac3d-9036-421f-aade-102c1737e38e-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:02:23 crc kubenswrapper[4824]: I1209 11:02:23.888275 4824 generic.go:334] "Generic (PLEG): container finished" podID="b350ac3d-9036-421f-aade-102c1737e38e" containerID="fc717ac53c7847e8ad1a1e98fc5066a0839ada0187706933e40d4b754c873d22" exitCode=0 Dec 09 11:02:23 crc kubenswrapper[4824]: I1209 11:02:23.888328 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5wvbr" Dec 09 11:02:23 crc kubenswrapper[4824]: I1209 11:02:23.888347 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5wvbr" event={"ID":"b350ac3d-9036-421f-aade-102c1737e38e","Type":"ContainerDied","Data":"fc717ac53c7847e8ad1a1e98fc5066a0839ada0187706933e40d4b754c873d22"} Dec 09 11:02:23 crc kubenswrapper[4824]: I1209 11:02:23.888704 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5wvbr" event={"ID":"b350ac3d-9036-421f-aade-102c1737e38e","Type":"ContainerDied","Data":"fb03ce3a34e295a2e7f2be9e451cf31899108b9a7b41900e8d90e1c11ac7b5d4"} Dec 09 11:02:23 crc kubenswrapper[4824]: I1209 11:02:23.888736 4824 scope.go:117] "RemoveContainer" containerID="fc717ac53c7847e8ad1a1e98fc5066a0839ada0187706933e40d4b754c873d22" Dec 09 11:02:23 crc kubenswrapper[4824]: I1209 11:02:23.933269 4824 scope.go:117] "RemoveContainer" containerID="3a9e3735e7d17f718166b9d859c00bb7ced6f5da1be983096fa9e8ce3565db66" Dec 09 11:02:23 crc kubenswrapper[4824]: I1209 11:02:23.956317 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5wvbr"] Dec 09 11:02:23 crc kubenswrapper[4824]: I1209 11:02:23.956356 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5wvbr"] Dec 09 11:02:23 crc kubenswrapper[4824]: I1209 11:02:23.961574 4824 scope.go:117] "RemoveContainer" containerID="b7a2c468892d2592e913149bc6bb4c054ca501c492d14d81213a818e78c12e87" Dec 09 11:02:24 crc kubenswrapper[4824]: I1209 11:02:24.022890 4824 scope.go:117] "RemoveContainer" containerID="fc717ac53c7847e8ad1a1e98fc5066a0839ada0187706933e40d4b754c873d22" Dec 09 11:02:24 crc kubenswrapper[4824]: E1209 11:02:24.023502 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc717ac53c7847e8ad1a1e98fc5066a0839ada0187706933e40d4b754c873d22\": container with ID starting with fc717ac53c7847e8ad1a1e98fc5066a0839ada0187706933e40d4b754c873d22 not found: ID does not exist" containerID="fc717ac53c7847e8ad1a1e98fc5066a0839ada0187706933e40d4b754c873d22" Dec 09 11:02:24 crc kubenswrapper[4824]: I1209 11:02:24.023574 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc717ac53c7847e8ad1a1e98fc5066a0839ada0187706933e40d4b754c873d22"} err="failed to get container status \"fc717ac53c7847e8ad1a1e98fc5066a0839ada0187706933e40d4b754c873d22\": rpc error: code = NotFound desc = could not find container \"fc717ac53c7847e8ad1a1e98fc5066a0839ada0187706933e40d4b754c873d22\": container with ID starting with fc717ac53c7847e8ad1a1e98fc5066a0839ada0187706933e40d4b754c873d22 not found: ID does not exist" Dec 09 11:02:24 crc kubenswrapper[4824]: I1209 11:02:24.023630 4824 scope.go:117] "RemoveContainer" containerID="3a9e3735e7d17f718166b9d859c00bb7ced6f5da1be983096fa9e8ce3565db66" Dec 09 11:02:24 crc kubenswrapper[4824]: E1209 11:02:24.024212 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a9e3735e7d17f718166b9d859c00bb7ced6f5da1be983096fa9e8ce3565db66\": container with ID starting with 3a9e3735e7d17f718166b9d859c00bb7ced6f5da1be983096fa9e8ce3565db66 not found: ID does not exist" containerID="3a9e3735e7d17f718166b9d859c00bb7ced6f5da1be983096fa9e8ce3565db66" Dec 09 11:02:24 crc kubenswrapper[4824]: I1209 11:02:24.024240 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a9e3735e7d17f718166b9d859c00bb7ced6f5da1be983096fa9e8ce3565db66"} err="failed to get container status \"3a9e3735e7d17f718166b9d859c00bb7ced6f5da1be983096fa9e8ce3565db66\": rpc error: code = NotFound desc = could not find container \"3a9e3735e7d17f718166b9d859c00bb7ced6f5da1be983096fa9e8ce3565db66\": container with ID starting with 3a9e3735e7d17f718166b9d859c00bb7ced6f5da1be983096fa9e8ce3565db66 not found: ID does not exist" Dec 09 11:02:24 crc kubenswrapper[4824]: I1209 11:02:24.024263 4824 scope.go:117] "RemoveContainer" containerID="b7a2c468892d2592e913149bc6bb4c054ca501c492d14d81213a818e78c12e87" Dec 09 11:02:24 crc kubenswrapper[4824]: E1209 11:02:24.024774 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7a2c468892d2592e913149bc6bb4c054ca501c492d14d81213a818e78c12e87\": container with ID starting with b7a2c468892d2592e913149bc6bb4c054ca501c492d14d81213a818e78c12e87 not found: ID does not exist" containerID="b7a2c468892d2592e913149bc6bb4c054ca501c492d14d81213a818e78c12e87" Dec 09 11:02:24 crc kubenswrapper[4824]: I1209 11:02:24.024854 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7a2c468892d2592e913149bc6bb4c054ca501c492d14d81213a818e78c12e87"} err="failed to get container status \"b7a2c468892d2592e913149bc6bb4c054ca501c492d14d81213a818e78c12e87\": rpc error: code = NotFound desc = could not find container \"b7a2c468892d2592e913149bc6bb4c054ca501c492d14d81213a818e78c12e87\": container with ID starting with b7a2c468892d2592e913149bc6bb4c054ca501c492d14d81213a818e78c12e87 not found: ID does not exist" Dec 09 11:02:24 crc kubenswrapper[4824]: E1209 11:02:24.426418 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb350ac3d_9036_421f_aade_102c1737e38e.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:02:25 crc kubenswrapper[4824]: I1209 11:02:25.926831 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b350ac3d-9036-421f-aade-102c1737e38e" path="/var/lib/kubelet/pods/b350ac3d-9036-421f-aade-102c1737e38e/volumes" Dec 09 11:02:32 crc kubenswrapper[4824]: I1209 11:02:32.861445 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:02:32 crc kubenswrapper[4824]: I1209 11:02:32.861878 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:02:34 crc kubenswrapper[4824]: E1209 11:02:34.739474 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb350ac3d_9036_421f_aade_102c1737e38e.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:02:38 crc kubenswrapper[4824]: E1209 11:02:38.241278 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb350ac3d_9036_421f_aade_102c1737e38e.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:02:45 crc kubenswrapper[4824]: E1209 11:02:45.136507 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb350ac3d_9036_421f_aade_102c1737e38e.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:02:48 crc kubenswrapper[4824]: E1209 11:02:48.127536 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb350ac3d_9036_421f_aade_102c1737e38e.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:02:48 crc kubenswrapper[4824]: E1209 11:02:48.128566 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb350ac3d_9036_421f_aade_102c1737e38e.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:02:53 crc kubenswrapper[4824]: E1209 11:02:53.525519 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb350ac3d_9036_421f_aade_102c1737e38e.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:02:55 crc kubenswrapper[4824]: E1209 11:02:55.179404 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb350ac3d_9036_421f_aade_102c1737e38e.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:03:02 crc kubenswrapper[4824]: I1209 11:03:02.860481 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:03:02 crc kubenswrapper[4824]: I1209 11:03:02.860969 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:03:05 crc kubenswrapper[4824]: E1209 11:03:05.229486 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb350ac3d_9036_421f_aade_102c1737e38e.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:03:08 crc kubenswrapper[4824]: E1209 11:03:08.531172 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb350ac3d_9036_421f_aade_102c1737e38e.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:03:15 crc kubenswrapper[4824]: E1209 11:03:15.558325 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb350ac3d_9036_421f_aade_102c1737e38e.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:03:23 crc kubenswrapper[4824]: E1209 11:03:23.501240 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb350ac3d_9036_421f_aade_102c1737e38e.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:03:32 crc kubenswrapper[4824]: I1209 11:03:32.861057 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:03:32 crc kubenswrapper[4824]: I1209 11:03:32.861562 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:03:32 crc kubenswrapper[4824]: I1209 11:03:32.861615 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 11:03:32 crc kubenswrapper[4824]: I1209 11:03:32.862678 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4"} pod="openshift-machine-config-operator/machine-config-daemon-dth8x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 11:03:32 crc kubenswrapper[4824]: I1209 11:03:32.862745 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" containerID="cri-o://2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" gracePeriod=600 Dec 09 11:03:33 crc kubenswrapper[4824]: E1209 11:03:33.492147 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:03:33 crc kubenswrapper[4824]: I1209 11:03:33.972476 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" exitCode=0 Dec 09 11:03:33 crc kubenswrapper[4824]: I1209 11:03:33.972537 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerDied","Data":"2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4"} Dec 09 11:03:33 crc kubenswrapper[4824]: I1209 11:03:33.972577 4824 scope.go:117] "RemoveContainer" containerID="c5001da7621385ad85d34d158a04bc1fc9f0d5d5dcdc6b091e37b7dda74aee0b" Dec 09 11:03:33 crc kubenswrapper[4824]: I1209 11:03:33.973532 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:03:33 crc kubenswrapper[4824]: E1209 11:03:33.974018 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:03:44 crc kubenswrapper[4824]: I1209 11:03:44.910308 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:03:44 crc kubenswrapper[4824]: E1209 11:03:44.911899 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:03:56 crc kubenswrapper[4824]: I1209 11:03:56.911710 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:03:56 crc kubenswrapper[4824]: E1209 11:03:56.912673 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:04:10 crc kubenswrapper[4824]: I1209 11:04:10.911141 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:04:10 crc kubenswrapper[4824]: E1209 11:04:10.911933 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:04:24 crc kubenswrapper[4824]: I1209 11:04:24.911321 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:04:24 crc kubenswrapper[4824]: E1209 11:04:24.912198 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:04:39 crc kubenswrapper[4824]: I1209 11:04:39.910866 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:04:39 crc kubenswrapper[4824]: E1209 11:04:39.912174 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:04:54 crc kubenswrapper[4824]: I1209 11:04:54.911178 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:04:54 crc kubenswrapper[4824]: E1209 11:04:54.912424 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:04:55 crc kubenswrapper[4824]: E1209 11:04:55.251272 4824 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.128:47042->38.102.83.128:33749: write tcp 38.102.83.128:47042->38.102.83.128:33749: write: broken pipe Dec 09 11:05:06 crc kubenswrapper[4824]: I1209 11:05:06.911794 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:05:06 crc kubenswrapper[4824]: E1209 11:05:06.913132 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:05:20 crc kubenswrapper[4824]: I1209 11:05:20.911715 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:05:20 crc kubenswrapper[4824]: E1209 11:05:20.912559 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:05:26 crc kubenswrapper[4824]: I1209 11:05:26.955995 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dllvn"] Dec 09 11:05:26 crc kubenswrapper[4824]: E1209 11:05:26.957235 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b350ac3d-9036-421f-aade-102c1737e38e" containerName="registry-server" Dec 09 11:05:26 crc kubenswrapper[4824]: I1209 11:05:26.957254 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b350ac3d-9036-421f-aade-102c1737e38e" containerName="registry-server" Dec 09 11:05:26 crc kubenswrapper[4824]: E1209 11:05:26.957325 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b350ac3d-9036-421f-aade-102c1737e38e" containerName="extract-content" Dec 09 11:05:26 crc kubenswrapper[4824]: I1209 11:05:26.957334 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b350ac3d-9036-421f-aade-102c1737e38e" containerName="extract-content" Dec 09 11:05:26 crc kubenswrapper[4824]: E1209 11:05:26.957354 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b350ac3d-9036-421f-aade-102c1737e38e" containerName="extract-utilities" Dec 09 11:05:26 crc kubenswrapper[4824]: I1209 11:05:26.957365 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b350ac3d-9036-421f-aade-102c1737e38e" containerName="extract-utilities" Dec 09 11:05:26 crc kubenswrapper[4824]: I1209 11:05:26.957679 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b350ac3d-9036-421f-aade-102c1737e38e" containerName="registry-server" Dec 09 11:05:26 crc kubenswrapper[4824]: I1209 11:05:26.969826 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dllvn" Dec 09 11:05:26 crc kubenswrapper[4824]: I1209 11:05:26.991007 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dllvn"] Dec 09 11:05:27 crc kubenswrapper[4824]: I1209 11:05:27.035484 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ddb37e7-b54a-40aa-85ee-b73aacb4910b-utilities\") pod \"redhat-operators-dllvn\" (UID: \"4ddb37e7-b54a-40aa-85ee-b73aacb4910b\") " pod="openshift-marketplace/redhat-operators-dllvn" Dec 09 11:05:27 crc kubenswrapper[4824]: I1209 11:05:27.035607 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ddb37e7-b54a-40aa-85ee-b73aacb4910b-catalog-content\") pod \"redhat-operators-dllvn\" (UID: \"4ddb37e7-b54a-40aa-85ee-b73aacb4910b\") " pod="openshift-marketplace/redhat-operators-dllvn" Dec 09 11:05:27 crc kubenswrapper[4824]: I1209 11:05:27.035959 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkznb\" (UniqueName: \"kubernetes.io/projected/4ddb37e7-b54a-40aa-85ee-b73aacb4910b-kube-api-access-zkznb\") pod \"redhat-operators-dllvn\" (UID: \"4ddb37e7-b54a-40aa-85ee-b73aacb4910b\") " pod="openshift-marketplace/redhat-operators-dllvn" Dec 09 11:05:27 crc kubenswrapper[4824]: I1209 11:05:27.137863 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkznb\" (UniqueName: \"kubernetes.io/projected/4ddb37e7-b54a-40aa-85ee-b73aacb4910b-kube-api-access-zkznb\") pod \"redhat-operators-dllvn\" (UID: \"4ddb37e7-b54a-40aa-85ee-b73aacb4910b\") " pod="openshift-marketplace/redhat-operators-dllvn" Dec 09 11:05:27 crc kubenswrapper[4824]: I1209 11:05:27.137947 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ddb37e7-b54a-40aa-85ee-b73aacb4910b-utilities\") pod \"redhat-operators-dllvn\" (UID: \"4ddb37e7-b54a-40aa-85ee-b73aacb4910b\") " pod="openshift-marketplace/redhat-operators-dllvn" Dec 09 11:05:27 crc kubenswrapper[4824]: I1209 11:05:27.138001 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ddb37e7-b54a-40aa-85ee-b73aacb4910b-catalog-content\") pod \"redhat-operators-dllvn\" (UID: \"4ddb37e7-b54a-40aa-85ee-b73aacb4910b\") " pod="openshift-marketplace/redhat-operators-dllvn" Dec 09 11:05:27 crc kubenswrapper[4824]: I1209 11:05:27.138652 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ddb37e7-b54a-40aa-85ee-b73aacb4910b-catalog-content\") pod \"redhat-operators-dllvn\" (UID: \"4ddb37e7-b54a-40aa-85ee-b73aacb4910b\") " pod="openshift-marketplace/redhat-operators-dllvn" Dec 09 11:05:27 crc kubenswrapper[4824]: I1209 11:05:27.138695 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ddb37e7-b54a-40aa-85ee-b73aacb4910b-utilities\") pod \"redhat-operators-dllvn\" (UID: \"4ddb37e7-b54a-40aa-85ee-b73aacb4910b\") " pod="openshift-marketplace/redhat-operators-dllvn" Dec 09 11:05:27 crc kubenswrapper[4824]: I1209 11:05:27.168543 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkznb\" (UniqueName: \"kubernetes.io/projected/4ddb37e7-b54a-40aa-85ee-b73aacb4910b-kube-api-access-zkznb\") pod \"redhat-operators-dllvn\" (UID: \"4ddb37e7-b54a-40aa-85ee-b73aacb4910b\") " pod="openshift-marketplace/redhat-operators-dllvn" Dec 09 11:05:27 crc kubenswrapper[4824]: I1209 11:05:27.327731 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dllvn" Dec 09 11:05:27 crc kubenswrapper[4824]: I1209 11:05:27.857863 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dllvn"] Dec 09 11:05:28 crc kubenswrapper[4824]: I1209 11:05:28.030977 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dllvn" event={"ID":"4ddb37e7-b54a-40aa-85ee-b73aacb4910b","Type":"ContainerStarted","Data":"a2a76eefe5e497a512a0feba0ceef2cf01679d03d49c7f6731deb68a2b1ae163"} Dec 09 11:05:29 crc kubenswrapper[4824]: I1209 11:05:29.044270 4824 generic.go:334] "Generic (PLEG): container finished" podID="4ddb37e7-b54a-40aa-85ee-b73aacb4910b" containerID="f989670a4f075eff116142594d44d4f939788f0a22b714bd212bc329f8970a72" exitCode=0 Dec 09 11:05:29 crc kubenswrapper[4824]: I1209 11:05:29.044318 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dllvn" event={"ID":"4ddb37e7-b54a-40aa-85ee-b73aacb4910b","Type":"ContainerDied","Data":"f989670a4f075eff116142594d44d4f939788f0a22b714bd212bc329f8970a72"} Dec 09 11:05:31 crc kubenswrapper[4824]: I1209 11:05:31.070719 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dllvn" event={"ID":"4ddb37e7-b54a-40aa-85ee-b73aacb4910b","Type":"ContainerStarted","Data":"c771247c8d68914da1a106aa8cbe4522417274bc74292e6f1ad5ec19c8eb3c29"} Dec 09 11:05:33 crc kubenswrapper[4824]: I1209 11:05:33.098774 4824 generic.go:334] "Generic (PLEG): container finished" podID="4ddb37e7-b54a-40aa-85ee-b73aacb4910b" containerID="c771247c8d68914da1a106aa8cbe4522417274bc74292e6f1ad5ec19c8eb3c29" exitCode=0 Dec 09 11:05:33 crc kubenswrapper[4824]: I1209 11:05:33.098908 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dllvn" event={"ID":"4ddb37e7-b54a-40aa-85ee-b73aacb4910b","Type":"ContainerDied","Data":"c771247c8d68914da1a106aa8cbe4522417274bc74292e6f1ad5ec19c8eb3c29"} Dec 09 11:05:35 crc kubenswrapper[4824]: I1209 11:05:35.129123 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dllvn" event={"ID":"4ddb37e7-b54a-40aa-85ee-b73aacb4910b","Type":"ContainerStarted","Data":"82eb829f3856ca9bd899d547053fa5363889c891546d81c2945185a954a52106"} Dec 09 11:05:35 crc kubenswrapper[4824]: I1209 11:05:35.162012 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dllvn" podStartSLOduration=4.683884966 podStartE2EDuration="9.161954108s" podCreationTimestamp="2025-12-09 11:05:26 +0000 UTC" firstStartedPulling="2025-12-09 11:05:29.046279435 +0000 UTC m=+4685.380784102" lastFinishedPulling="2025-12-09 11:05:33.524348577 +0000 UTC m=+4689.858853244" observedRunningTime="2025-12-09 11:05:35.152928575 +0000 UTC m=+4691.487433242" watchObservedRunningTime="2025-12-09 11:05:35.161954108 +0000 UTC m=+4691.496458805" Dec 09 11:05:35 crc kubenswrapper[4824]: I1209 11:05:35.910732 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:05:35 crc kubenswrapper[4824]: E1209 11:05:35.911463 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:05:37 crc kubenswrapper[4824]: I1209 11:05:37.328320 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dllvn" Dec 09 11:05:37 crc kubenswrapper[4824]: I1209 11:05:37.328634 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dllvn" Dec 09 11:05:38 crc kubenswrapper[4824]: I1209 11:05:38.395602 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dllvn" podUID="4ddb37e7-b54a-40aa-85ee-b73aacb4910b" containerName="registry-server" probeResult="failure" output=< Dec 09 11:05:38 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 11:05:38 crc kubenswrapper[4824]: > Dec 09 11:05:48 crc kubenswrapper[4824]: I1209 11:05:48.538226 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dllvn" podUID="4ddb37e7-b54a-40aa-85ee-b73aacb4910b" containerName="registry-server" probeResult="failure" output=< Dec 09 11:05:48 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 11:05:48 crc kubenswrapper[4824]: > Dec 09 11:05:48 crc kubenswrapper[4824]: I1209 11:05:48.911597 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:05:48 crc kubenswrapper[4824]: E1209 11:05:48.912131 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:05:57 crc kubenswrapper[4824]: I1209 11:05:57.793764 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dllvn" Dec 09 11:05:57 crc kubenswrapper[4824]: I1209 11:05:57.854582 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dllvn" Dec 09 11:05:58 crc kubenswrapper[4824]: I1209 11:05:58.134604 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dllvn"] Dec 09 11:05:59 crc kubenswrapper[4824]: I1209 11:05:59.755708 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dllvn" podUID="4ddb37e7-b54a-40aa-85ee-b73aacb4910b" containerName="registry-server" containerID="cri-o://82eb829f3856ca9bd899d547053fa5363889c891546d81c2945185a954a52106" gracePeriod=2 Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.543615 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dllvn" Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.613202 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ddb37e7-b54a-40aa-85ee-b73aacb4910b-catalog-content\") pod \"4ddb37e7-b54a-40aa-85ee-b73aacb4910b\" (UID: \"4ddb37e7-b54a-40aa-85ee-b73aacb4910b\") " Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.613320 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ddb37e7-b54a-40aa-85ee-b73aacb4910b-utilities\") pod \"4ddb37e7-b54a-40aa-85ee-b73aacb4910b\" (UID: \"4ddb37e7-b54a-40aa-85ee-b73aacb4910b\") " Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.613541 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkznb\" (UniqueName: \"kubernetes.io/projected/4ddb37e7-b54a-40aa-85ee-b73aacb4910b-kube-api-access-zkznb\") pod \"4ddb37e7-b54a-40aa-85ee-b73aacb4910b\" (UID: \"4ddb37e7-b54a-40aa-85ee-b73aacb4910b\") " Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.614462 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ddb37e7-b54a-40aa-85ee-b73aacb4910b-utilities" (OuterVolumeSpecName: "utilities") pod "4ddb37e7-b54a-40aa-85ee-b73aacb4910b" (UID: "4ddb37e7-b54a-40aa-85ee-b73aacb4910b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.622051 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ddb37e7-b54a-40aa-85ee-b73aacb4910b-kube-api-access-zkznb" (OuterVolumeSpecName: "kube-api-access-zkznb") pod "4ddb37e7-b54a-40aa-85ee-b73aacb4910b" (UID: "4ddb37e7-b54a-40aa-85ee-b73aacb4910b"). InnerVolumeSpecName "kube-api-access-zkznb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.717193 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ddb37e7-b54a-40aa-85ee-b73aacb4910b-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.717228 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkznb\" (UniqueName: \"kubernetes.io/projected/4ddb37e7-b54a-40aa-85ee-b73aacb4910b-kube-api-access-zkznb\") on node \"crc\" DevicePath \"\"" Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.721692 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ddb37e7-b54a-40aa-85ee-b73aacb4910b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4ddb37e7-b54a-40aa-85ee-b73aacb4910b" (UID: "4ddb37e7-b54a-40aa-85ee-b73aacb4910b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.767849 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dllvn" Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.767846 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dllvn" event={"ID":"4ddb37e7-b54a-40aa-85ee-b73aacb4910b","Type":"ContainerDied","Data":"82eb829f3856ca9bd899d547053fa5363889c891546d81c2945185a954a52106"} Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.767946 4824 scope.go:117] "RemoveContainer" containerID="82eb829f3856ca9bd899d547053fa5363889c891546d81c2945185a954a52106" Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.767770 4824 generic.go:334] "Generic (PLEG): container finished" podID="4ddb37e7-b54a-40aa-85ee-b73aacb4910b" containerID="82eb829f3856ca9bd899d547053fa5363889c891546d81c2945185a954a52106" exitCode=0 Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.768050 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dllvn" event={"ID":"4ddb37e7-b54a-40aa-85ee-b73aacb4910b","Type":"ContainerDied","Data":"a2a76eefe5e497a512a0feba0ceef2cf01679d03d49c7f6731deb68a2b1ae163"} Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.803088 4824 scope.go:117] "RemoveContainer" containerID="c771247c8d68914da1a106aa8cbe4522417274bc74292e6f1ad5ec19c8eb3c29" Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.813202 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dllvn"] Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.820719 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ddb37e7-b54a-40aa-85ee-b73aacb4910b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.826644 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dllvn"] Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.828087 4824 scope.go:117] "RemoveContainer" containerID="f989670a4f075eff116142594d44d4f939788f0a22b714bd212bc329f8970a72" Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.899526 4824 scope.go:117] "RemoveContainer" containerID="82eb829f3856ca9bd899d547053fa5363889c891546d81c2945185a954a52106" Dec 09 11:06:00 crc kubenswrapper[4824]: E1209 11:06:00.900054 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82eb829f3856ca9bd899d547053fa5363889c891546d81c2945185a954a52106\": container with ID starting with 82eb829f3856ca9bd899d547053fa5363889c891546d81c2945185a954a52106 not found: ID does not exist" containerID="82eb829f3856ca9bd899d547053fa5363889c891546d81c2945185a954a52106" Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.900092 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82eb829f3856ca9bd899d547053fa5363889c891546d81c2945185a954a52106"} err="failed to get container status \"82eb829f3856ca9bd899d547053fa5363889c891546d81c2945185a954a52106\": rpc error: code = NotFound desc = could not find container \"82eb829f3856ca9bd899d547053fa5363889c891546d81c2945185a954a52106\": container with ID starting with 82eb829f3856ca9bd899d547053fa5363889c891546d81c2945185a954a52106 not found: ID does not exist" Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.900127 4824 scope.go:117] "RemoveContainer" containerID="c771247c8d68914da1a106aa8cbe4522417274bc74292e6f1ad5ec19c8eb3c29" Dec 09 11:06:00 crc kubenswrapper[4824]: E1209 11:06:00.900365 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c771247c8d68914da1a106aa8cbe4522417274bc74292e6f1ad5ec19c8eb3c29\": container with ID starting with c771247c8d68914da1a106aa8cbe4522417274bc74292e6f1ad5ec19c8eb3c29 not found: ID does not exist" containerID="c771247c8d68914da1a106aa8cbe4522417274bc74292e6f1ad5ec19c8eb3c29" Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.900389 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c771247c8d68914da1a106aa8cbe4522417274bc74292e6f1ad5ec19c8eb3c29"} err="failed to get container status \"c771247c8d68914da1a106aa8cbe4522417274bc74292e6f1ad5ec19c8eb3c29\": rpc error: code = NotFound desc = could not find container \"c771247c8d68914da1a106aa8cbe4522417274bc74292e6f1ad5ec19c8eb3c29\": container with ID starting with c771247c8d68914da1a106aa8cbe4522417274bc74292e6f1ad5ec19c8eb3c29 not found: ID does not exist" Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.900403 4824 scope.go:117] "RemoveContainer" containerID="f989670a4f075eff116142594d44d4f939788f0a22b714bd212bc329f8970a72" Dec 09 11:06:00 crc kubenswrapper[4824]: E1209 11:06:00.900577 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f989670a4f075eff116142594d44d4f939788f0a22b714bd212bc329f8970a72\": container with ID starting with f989670a4f075eff116142594d44d4f939788f0a22b714bd212bc329f8970a72 not found: ID does not exist" containerID="f989670a4f075eff116142594d44d4f939788f0a22b714bd212bc329f8970a72" Dec 09 11:06:00 crc kubenswrapper[4824]: I1209 11:06:00.900598 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f989670a4f075eff116142594d44d4f939788f0a22b714bd212bc329f8970a72"} err="failed to get container status \"f989670a4f075eff116142594d44d4f939788f0a22b714bd212bc329f8970a72\": rpc error: code = NotFound desc = could not find container \"f989670a4f075eff116142594d44d4f939788f0a22b714bd212bc329f8970a72\": container with ID starting with f989670a4f075eff116142594d44d4f939788f0a22b714bd212bc329f8970a72 not found: ID does not exist" Dec 09 11:06:01 crc kubenswrapper[4824]: I1209 11:06:01.925043 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ddb37e7-b54a-40aa-85ee-b73aacb4910b" path="/var/lib/kubelet/pods/4ddb37e7-b54a-40aa-85ee-b73aacb4910b/volumes" Dec 09 11:06:02 crc kubenswrapper[4824]: I1209 11:06:02.911914 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:06:02 crc kubenswrapper[4824]: E1209 11:06:02.912692 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:06:16 crc kubenswrapper[4824]: I1209 11:06:16.910280 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:06:16 crc kubenswrapper[4824]: E1209 11:06:16.911152 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:06:25 crc kubenswrapper[4824]: E1209 11:06:25.520523 4824 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.128:49840->38.102.83.128:33749: write tcp 38.102.83.128:49840->38.102.83.128:33749: write: broken pipe Dec 09 11:06:30 crc kubenswrapper[4824]: I1209 11:06:30.912438 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:06:30 crc kubenswrapper[4824]: E1209 11:06:30.913225 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:06:41 crc kubenswrapper[4824]: I1209 11:06:41.910764 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:06:41 crc kubenswrapper[4824]: E1209 11:06:41.911616 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:06:55 crc kubenswrapper[4824]: I1209 11:06:55.911553 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:06:55 crc kubenswrapper[4824]: E1209 11:06:55.912864 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:07:08 crc kubenswrapper[4824]: I1209 11:07:08.912065 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:07:08 crc kubenswrapper[4824]: E1209 11:07:08.912908 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:07:12 crc kubenswrapper[4824]: I1209 11:07:12.692439 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zkgt7"] Dec 09 11:07:12 crc kubenswrapper[4824]: E1209 11:07:12.694216 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ddb37e7-b54a-40aa-85ee-b73aacb4910b" containerName="extract-utilities" Dec 09 11:07:12 crc kubenswrapper[4824]: I1209 11:07:12.694241 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ddb37e7-b54a-40aa-85ee-b73aacb4910b" containerName="extract-utilities" Dec 09 11:07:12 crc kubenswrapper[4824]: E1209 11:07:12.694298 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ddb37e7-b54a-40aa-85ee-b73aacb4910b" containerName="registry-server" Dec 09 11:07:12 crc kubenswrapper[4824]: I1209 11:07:12.694316 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ddb37e7-b54a-40aa-85ee-b73aacb4910b" containerName="registry-server" Dec 09 11:07:12 crc kubenswrapper[4824]: E1209 11:07:12.694348 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ddb37e7-b54a-40aa-85ee-b73aacb4910b" containerName="extract-content" Dec 09 11:07:12 crc kubenswrapper[4824]: I1209 11:07:12.694364 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ddb37e7-b54a-40aa-85ee-b73aacb4910b" containerName="extract-content" Dec 09 11:07:12 crc kubenswrapper[4824]: I1209 11:07:12.695112 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ddb37e7-b54a-40aa-85ee-b73aacb4910b" containerName="registry-server" Dec 09 11:07:12 crc kubenswrapper[4824]: I1209 11:07:12.699023 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zkgt7" Dec 09 11:07:12 crc kubenswrapper[4824]: I1209 11:07:12.712209 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zkgt7"] Dec 09 11:07:12 crc kubenswrapper[4824]: I1209 11:07:12.889078 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f0fac7c-e2a6-449a-ac69-ae4812cd722b-catalog-content\") pod \"community-operators-zkgt7\" (UID: \"5f0fac7c-e2a6-449a-ac69-ae4812cd722b\") " pod="openshift-marketplace/community-operators-zkgt7" Dec 09 11:07:12 crc kubenswrapper[4824]: I1209 11:07:12.889275 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w428\" (UniqueName: \"kubernetes.io/projected/5f0fac7c-e2a6-449a-ac69-ae4812cd722b-kube-api-access-9w428\") pod \"community-operators-zkgt7\" (UID: \"5f0fac7c-e2a6-449a-ac69-ae4812cd722b\") " pod="openshift-marketplace/community-operators-zkgt7" Dec 09 11:07:12 crc kubenswrapper[4824]: I1209 11:07:12.889308 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f0fac7c-e2a6-449a-ac69-ae4812cd722b-utilities\") pod \"community-operators-zkgt7\" (UID: \"5f0fac7c-e2a6-449a-ac69-ae4812cd722b\") " pod="openshift-marketplace/community-operators-zkgt7" Dec 09 11:07:12 crc kubenswrapper[4824]: I1209 11:07:12.991895 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f0fac7c-e2a6-449a-ac69-ae4812cd722b-catalog-content\") pod \"community-operators-zkgt7\" (UID: \"5f0fac7c-e2a6-449a-ac69-ae4812cd722b\") " pod="openshift-marketplace/community-operators-zkgt7" Dec 09 11:07:12 crc kubenswrapper[4824]: I1209 11:07:12.992110 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w428\" (UniqueName: \"kubernetes.io/projected/5f0fac7c-e2a6-449a-ac69-ae4812cd722b-kube-api-access-9w428\") pod \"community-operators-zkgt7\" (UID: \"5f0fac7c-e2a6-449a-ac69-ae4812cd722b\") " pod="openshift-marketplace/community-operators-zkgt7" Dec 09 11:07:12 crc kubenswrapper[4824]: I1209 11:07:12.992147 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f0fac7c-e2a6-449a-ac69-ae4812cd722b-utilities\") pod \"community-operators-zkgt7\" (UID: \"5f0fac7c-e2a6-449a-ac69-ae4812cd722b\") " pod="openshift-marketplace/community-operators-zkgt7" Dec 09 11:07:12 crc kubenswrapper[4824]: I1209 11:07:12.992577 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f0fac7c-e2a6-449a-ac69-ae4812cd722b-catalog-content\") pod \"community-operators-zkgt7\" (UID: \"5f0fac7c-e2a6-449a-ac69-ae4812cd722b\") " pod="openshift-marketplace/community-operators-zkgt7" Dec 09 11:07:12 crc kubenswrapper[4824]: I1209 11:07:12.992776 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f0fac7c-e2a6-449a-ac69-ae4812cd722b-utilities\") pod \"community-operators-zkgt7\" (UID: \"5f0fac7c-e2a6-449a-ac69-ae4812cd722b\") " pod="openshift-marketplace/community-operators-zkgt7" Dec 09 11:07:13 crc kubenswrapper[4824]: I1209 11:07:13.014879 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w428\" (UniqueName: \"kubernetes.io/projected/5f0fac7c-e2a6-449a-ac69-ae4812cd722b-kube-api-access-9w428\") pod \"community-operators-zkgt7\" (UID: \"5f0fac7c-e2a6-449a-ac69-ae4812cd722b\") " pod="openshift-marketplace/community-operators-zkgt7" Dec 09 11:07:13 crc kubenswrapper[4824]: I1209 11:07:13.032716 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zkgt7" Dec 09 11:07:13 crc kubenswrapper[4824]: I1209 11:07:13.594222 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zkgt7"] Dec 09 11:07:13 crc kubenswrapper[4824]: W1209 11:07:13.607952 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f0fac7c_e2a6_449a_ac69_ae4812cd722b.slice/crio-2ac7d75e4dac67f568fc54730d8e51317dc053b0ca723ff84150639f1ce17a4d WatchSource:0}: Error finding container 2ac7d75e4dac67f568fc54730d8e51317dc053b0ca723ff84150639f1ce17a4d: Status 404 returned error can't find the container with id 2ac7d75e4dac67f568fc54730d8e51317dc053b0ca723ff84150639f1ce17a4d Dec 09 11:07:14 crc kubenswrapper[4824]: I1209 11:07:14.100856 4824 generic.go:334] "Generic (PLEG): container finished" podID="5f0fac7c-e2a6-449a-ac69-ae4812cd722b" containerID="51b363b5177baf078d73d1a04071d4204ec6bd1bc41625e22e7e36e93966e047" exitCode=0 Dec 09 11:07:14 crc kubenswrapper[4824]: I1209 11:07:14.101138 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zkgt7" event={"ID":"5f0fac7c-e2a6-449a-ac69-ae4812cd722b","Type":"ContainerDied","Data":"51b363b5177baf078d73d1a04071d4204ec6bd1bc41625e22e7e36e93966e047"} Dec 09 11:07:14 crc kubenswrapper[4824]: I1209 11:07:14.101166 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zkgt7" event={"ID":"5f0fac7c-e2a6-449a-ac69-ae4812cd722b","Type":"ContainerStarted","Data":"2ac7d75e4dac67f568fc54730d8e51317dc053b0ca723ff84150639f1ce17a4d"} Dec 09 11:07:14 crc kubenswrapper[4824]: I1209 11:07:14.103817 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 11:07:18 crc kubenswrapper[4824]: I1209 11:07:18.149873 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zkgt7" event={"ID":"5f0fac7c-e2a6-449a-ac69-ae4812cd722b","Type":"ContainerStarted","Data":"505537358d0f511921e7382bafac0d5d8fcbdda64f47520430663e0ff8df87a7"} Dec 09 11:07:19 crc kubenswrapper[4824]: I1209 11:07:19.163523 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zkgt7" event={"ID":"5f0fac7c-e2a6-449a-ac69-ae4812cd722b","Type":"ContainerDied","Data":"505537358d0f511921e7382bafac0d5d8fcbdda64f47520430663e0ff8df87a7"} Dec 09 11:07:19 crc kubenswrapper[4824]: I1209 11:07:19.163363 4824 generic.go:334] "Generic (PLEG): container finished" podID="5f0fac7c-e2a6-449a-ac69-ae4812cd722b" containerID="505537358d0f511921e7382bafac0d5d8fcbdda64f47520430663e0ff8df87a7" exitCode=0 Dec 09 11:07:22 crc kubenswrapper[4824]: I1209 11:07:22.213356 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zkgt7" event={"ID":"5f0fac7c-e2a6-449a-ac69-ae4812cd722b","Type":"ContainerStarted","Data":"c5148846374248316a5116ed6e24f61894a857554b8d3520b518a504ae664fff"} Dec 09 11:07:22 crc kubenswrapper[4824]: I1209 11:07:22.237121 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zkgt7" podStartSLOduration=3.914831569 podStartE2EDuration="10.237102692s" podCreationTimestamp="2025-12-09 11:07:12 +0000 UTC" firstStartedPulling="2025-12-09 11:07:14.103573803 +0000 UTC m=+4790.438078470" lastFinishedPulling="2025-12-09 11:07:20.425844926 +0000 UTC m=+4796.760349593" observedRunningTime="2025-12-09 11:07:22.235348747 +0000 UTC m=+4798.569853424" watchObservedRunningTime="2025-12-09 11:07:22.237102692 +0000 UTC m=+4798.571607359" Dec 09 11:07:23 crc kubenswrapper[4824]: I1209 11:07:23.033116 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zkgt7" Dec 09 11:07:23 crc kubenswrapper[4824]: I1209 11:07:23.033225 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zkgt7" Dec 09 11:07:23 crc kubenswrapper[4824]: I1209 11:07:23.938557 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:07:23 crc kubenswrapper[4824]: E1209 11:07:23.950421 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:07:24 crc kubenswrapper[4824]: I1209 11:07:24.113388 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-zkgt7" podUID="5f0fac7c-e2a6-449a-ac69-ae4812cd722b" containerName="registry-server" probeResult="failure" output=< Dec 09 11:07:24 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 11:07:24 crc kubenswrapper[4824]: > Dec 09 11:07:33 crc kubenswrapper[4824]: I1209 11:07:33.284702 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zkgt7" Dec 09 11:07:33 crc kubenswrapper[4824]: I1209 11:07:33.342289 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zkgt7" Dec 09 11:07:33 crc kubenswrapper[4824]: I1209 11:07:33.521503 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zkgt7"] Dec 09 11:07:34 crc kubenswrapper[4824]: I1209 11:07:34.357873 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zkgt7" podUID="5f0fac7c-e2a6-449a-ac69-ae4812cd722b" containerName="registry-server" containerID="cri-o://c5148846374248316a5116ed6e24f61894a857554b8d3520b518a504ae664fff" gracePeriod=2 Dec 09 11:07:34 crc kubenswrapper[4824]: I1209 11:07:34.923342 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zkgt7" Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.022847 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f0fac7c-e2a6-449a-ac69-ae4812cd722b-utilities\") pod \"5f0fac7c-e2a6-449a-ac69-ae4812cd722b\" (UID: \"5f0fac7c-e2a6-449a-ac69-ae4812cd722b\") " Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.023163 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f0fac7c-e2a6-449a-ac69-ae4812cd722b-catalog-content\") pod \"5f0fac7c-e2a6-449a-ac69-ae4812cd722b\" (UID: \"5f0fac7c-e2a6-449a-ac69-ae4812cd722b\") " Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.023382 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9w428\" (UniqueName: \"kubernetes.io/projected/5f0fac7c-e2a6-449a-ac69-ae4812cd722b-kube-api-access-9w428\") pod \"5f0fac7c-e2a6-449a-ac69-ae4812cd722b\" (UID: \"5f0fac7c-e2a6-449a-ac69-ae4812cd722b\") " Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.023860 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f0fac7c-e2a6-449a-ac69-ae4812cd722b-utilities" (OuterVolumeSpecName: "utilities") pod "5f0fac7c-e2a6-449a-ac69-ae4812cd722b" (UID: "5f0fac7c-e2a6-449a-ac69-ae4812cd722b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.024346 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f0fac7c-e2a6-449a-ac69-ae4812cd722b-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.032253 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f0fac7c-e2a6-449a-ac69-ae4812cd722b-kube-api-access-9w428" (OuterVolumeSpecName: "kube-api-access-9w428") pod "5f0fac7c-e2a6-449a-ac69-ae4812cd722b" (UID: "5f0fac7c-e2a6-449a-ac69-ae4812cd722b"). InnerVolumeSpecName "kube-api-access-9w428". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.075832 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f0fac7c-e2a6-449a-ac69-ae4812cd722b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5f0fac7c-e2a6-449a-ac69-ae4812cd722b" (UID: "5f0fac7c-e2a6-449a-ac69-ae4812cd722b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.128703 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9w428\" (UniqueName: \"kubernetes.io/projected/5f0fac7c-e2a6-449a-ac69-ae4812cd722b-kube-api-access-9w428\") on node \"crc\" DevicePath \"\"" Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.129391 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f0fac7c-e2a6-449a-ac69-ae4812cd722b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.377043 4824 generic.go:334] "Generic (PLEG): container finished" podID="5f0fac7c-e2a6-449a-ac69-ae4812cd722b" containerID="c5148846374248316a5116ed6e24f61894a857554b8d3520b518a504ae664fff" exitCode=0 Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.377095 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zkgt7" event={"ID":"5f0fac7c-e2a6-449a-ac69-ae4812cd722b","Type":"ContainerDied","Data":"c5148846374248316a5116ed6e24f61894a857554b8d3520b518a504ae664fff"} Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.377132 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zkgt7" event={"ID":"5f0fac7c-e2a6-449a-ac69-ae4812cd722b","Type":"ContainerDied","Data":"2ac7d75e4dac67f568fc54730d8e51317dc053b0ca723ff84150639f1ce17a4d"} Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.377160 4824 scope.go:117] "RemoveContainer" containerID="c5148846374248316a5116ed6e24f61894a857554b8d3520b518a504ae664fff" Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.377164 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zkgt7" Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.415118 4824 scope.go:117] "RemoveContainer" containerID="505537358d0f511921e7382bafac0d5d8fcbdda64f47520430663e0ff8df87a7" Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.425848 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zkgt7"] Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.434690 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zkgt7"] Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.448170 4824 scope.go:117] "RemoveContainer" containerID="51b363b5177baf078d73d1a04071d4204ec6bd1bc41625e22e7e36e93966e047" Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.502663 4824 scope.go:117] "RemoveContainer" containerID="c5148846374248316a5116ed6e24f61894a857554b8d3520b518a504ae664fff" Dec 09 11:07:35 crc kubenswrapper[4824]: E1209 11:07:35.503376 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5148846374248316a5116ed6e24f61894a857554b8d3520b518a504ae664fff\": container with ID starting with c5148846374248316a5116ed6e24f61894a857554b8d3520b518a504ae664fff not found: ID does not exist" containerID="c5148846374248316a5116ed6e24f61894a857554b8d3520b518a504ae664fff" Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.503426 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5148846374248316a5116ed6e24f61894a857554b8d3520b518a504ae664fff"} err="failed to get container status \"c5148846374248316a5116ed6e24f61894a857554b8d3520b518a504ae664fff\": rpc error: code = NotFound desc = could not find container \"c5148846374248316a5116ed6e24f61894a857554b8d3520b518a504ae664fff\": container with ID starting with c5148846374248316a5116ed6e24f61894a857554b8d3520b518a504ae664fff not found: ID does not exist" Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.503460 4824 scope.go:117] "RemoveContainer" containerID="505537358d0f511921e7382bafac0d5d8fcbdda64f47520430663e0ff8df87a7" Dec 09 11:07:35 crc kubenswrapper[4824]: E1209 11:07:35.503903 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"505537358d0f511921e7382bafac0d5d8fcbdda64f47520430663e0ff8df87a7\": container with ID starting with 505537358d0f511921e7382bafac0d5d8fcbdda64f47520430663e0ff8df87a7 not found: ID does not exist" containerID="505537358d0f511921e7382bafac0d5d8fcbdda64f47520430663e0ff8df87a7" Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.503925 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"505537358d0f511921e7382bafac0d5d8fcbdda64f47520430663e0ff8df87a7"} err="failed to get container status \"505537358d0f511921e7382bafac0d5d8fcbdda64f47520430663e0ff8df87a7\": rpc error: code = NotFound desc = could not find container \"505537358d0f511921e7382bafac0d5d8fcbdda64f47520430663e0ff8df87a7\": container with ID starting with 505537358d0f511921e7382bafac0d5d8fcbdda64f47520430663e0ff8df87a7 not found: ID does not exist" Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.503938 4824 scope.go:117] "RemoveContainer" containerID="51b363b5177baf078d73d1a04071d4204ec6bd1bc41625e22e7e36e93966e047" Dec 09 11:07:35 crc kubenswrapper[4824]: E1209 11:07:35.504960 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51b363b5177baf078d73d1a04071d4204ec6bd1bc41625e22e7e36e93966e047\": container with ID starting with 51b363b5177baf078d73d1a04071d4204ec6bd1bc41625e22e7e36e93966e047 not found: ID does not exist" containerID="51b363b5177baf078d73d1a04071d4204ec6bd1bc41625e22e7e36e93966e047" Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.504997 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51b363b5177baf078d73d1a04071d4204ec6bd1bc41625e22e7e36e93966e047"} err="failed to get container status \"51b363b5177baf078d73d1a04071d4204ec6bd1bc41625e22e7e36e93966e047\": rpc error: code = NotFound desc = could not find container \"51b363b5177baf078d73d1a04071d4204ec6bd1bc41625e22e7e36e93966e047\": container with ID starting with 51b363b5177baf078d73d1a04071d4204ec6bd1bc41625e22e7e36e93966e047 not found: ID does not exist" Dec 09 11:07:35 crc kubenswrapper[4824]: I1209 11:07:35.926706 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f0fac7c-e2a6-449a-ac69-ae4812cd722b" path="/var/lib/kubelet/pods/5f0fac7c-e2a6-449a-ac69-ae4812cd722b/volumes" Dec 09 11:07:38 crc kubenswrapper[4824]: I1209 11:07:38.910736 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:07:38 crc kubenswrapper[4824]: E1209 11:07:38.911385 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:07:52 crc kubenswrapper[4824]: I1209 11:07:52.911074 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:07:52 crc kubenswrapper[4824]: E1209 11:07:52.912112 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:08:03 crc kubenswrapper[4824]: I1209 11:08:03.923709 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:08:03 crc kubenswrapper[4824]: E1209 11:08:03.925726 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:08:15 crc kubenswrapper[4824]: I1209 11:08:15.911819 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:08:15 crc kubenswrapper[4824]: E1209 11:08:15.914186 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:08:28 crc kubenswrapper[4824]: I1209 11:08:28.911249 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:08:28 crc kubenswrapper[4824]: E1209 11:08:28.912287 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:08:42 crc kubenswrapper[4824]: I1209 11:08:42.911994 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:08:43 crc kubenswrapper[4824]: I1209 11:08:43.482646 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerStarted","Data":"a341fa1c8d0134458a190647787e775a46a7ce784fbe98c35ff532671dc5ea08"} Dec 09 11:11:02 crc kubenswrapper[4824]: I1209 11:11:02.861677 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:11:02 crc kubenswrapper[4824]: I1209 11:11:02.862188 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:11:32 crc kubenswrapper[4824]: I1209 11:11:32.861027 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:11:32 crc kubenswrapper[4824]: I1209 11:11:32.861813 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.274956 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 09 11:11:43 crc kubenswrapper[4824]: E1209 11:11:43.276376 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f0fac7c-e2a6-449a-ac69-ae4812cd722b" containerName="extract-content" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.276405 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f0fac7c-e2a6-449a-ac69-ae4812cd722b" containerName="extract-content" Dec 09 11:11:43 crc kubenswrapper[4824]: E1209 11:11:43.276424 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f0fac7c-e2a6-449a-ac69-ae4812cd722b" containerName="registry-server" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.276432 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f0fac7c-e2a6-449a-ac69-ae4812cd722b" containerName="registry-server" Dec 09 11:11:43 crc kubenswrapper[4824]: E1209 11:11:43.276503 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f0fac7c-e2a6-449a-ac69-ae4812cd722b" containerName="extract-utilities" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.276514 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f0fac7c-e2a6-449a-ac69-ae4812cd722b" containerName="extract-utilities" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.277000 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f0fac7c-e2a6-449a-ac69-ae4812cd722b" containerName="registry-server" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.280479 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.283401 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.283657 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-n76qc" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.283869 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.284748 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.302294 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.414170 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzfmv\" (UniqueName: \"kubernetes.io/projected/0fd5955c-4039-42a1-8ecd-0256bb683e61-kube-api-access-dzfmv\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.414234 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0fd5955c-4039-42a1-8ecd-0256bb683e61-config-data\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.414323 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0fd5955c-4039-42a1-8ecd-0256bb683e61-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.414368 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0fd5955c-4039-42a1-8ecd-0256bb683e61-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.414404 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fd5955c-4039-42a1-8ecd-0256bb683e61-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.414543 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0fd5955c-4039-42a1-8ecd-0256bb683e61-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.414573 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0fd5955c-4039-42a1-8ecd-0256bb683e61-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.414605 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0fd5955c-4039-42a1-8ecd-0256bb683e61-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.414637 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.518179 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzfmv\" (UniqueName: \"kubernetes.io/projected/0fd5955c-4039-42a1-8ecd-0256bb683e61-kube-api-access-dzfmv\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.518347 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0fd5955c-4039-42a1-8ecd-0256bb683e61-config-data\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.518518 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0fd5955c-4039-42a1-8ecd-0256bb683e61-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.518632 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0fd5955c-4039-42a1-8ecd-0256bb683e61-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.518746 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fd5955c-4039-42a1-8ecd-0256bb683e61-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.519027 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0fd5955c-4039-42a1-8ecd-0256bb683e61-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.519081 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0fd5955c-4039-42a1-8ecd-0256bb683e61-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.519131 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0fd5955c-4039-42a1-8ecd-0256bb683e61-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.519170 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.519577 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0fd5955c-4039-42a1-8ecd-0256bb683e61-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.519668 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0fd5955c-4039-42a1-8ecd-0256bb683e61-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.520217 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0fd5955c-4039-42a1-8ecd-0256bb683e61-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.520346 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.520527 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0fd5955c-4039-42a1-8ecd-0256bb683e61-config-data\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.528343 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0fd5955c-4039-42a1-8ecd-0256bb683e61-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.537620 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0fd5955c-4039-42a1-8ecd-0256bb683e61-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.538234 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fd5955c-4039-42a1-8ecd-0256bb683e61-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.540156 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzfmv\" (UniqueName: \"kubernetes.io/projected/0fd5955c-4039-42a1-8ecd-0256bb683e61-kube-api-access-dzfmv\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.570486 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " pod="openstack/tempest-tests-tempest" Dec 09 11:11:43 crc kubenswrapper[4824]: I1209 11:11:43.613856 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 09 11:11:44 crc kubenswrapper[4824]: I1209 11:11:44.211655 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 09 11:11:44 crc kubenswrapper[4824]: I1209 11:11:44.435588 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"0fd5955c-4039-42a1-8ecd-0256bb683e61","Type":"ContainerStarted","Data":"9f448c061e67b0ec27df28f2f694dbdbdce723f631d1c133ba2464a33b6531c5"} Dec 09 11:12:02 crc kubenswrapper[4824]: I1209 11:12:02.860636 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:12:02 crc kubenswrapper[4824]: I1209 11:12:02.861213 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:12:02 crc kubenswrapper[4824]: I1209 11:12:02.861266 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 11:12:02 crc kubenswrapper[4824]: I1209 11:12:02.862310 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a341fa1c8d0134458a190647787e775a46a7ce784fbe98c35ff532671dc5ea08"} pod="openshift-machine-config-operator/machine-config-daemon-dth8x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 11:12:02 crc kubenswrapper[4824]: I1209 11:12:02.862390 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" containerID="cri-o://a341fa1c8d0134458a190647787e775a46a7ce784fbe98c35ff532671dc5ea08" gracePeriod=600 Dec 09 11:12:03 crc kubenswrapper[4824]: I1209 11:12:03.708671 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerID="a341fa1c8d0134458a190647787e775a46a7ce784fbe98c35ff532671dc5ea08" exitCode=0 Dec 09 11:12:03 crc kubenswrapper[4824]: I1209 11:12:03.708717 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerDied","Data":"a341fa1c8d0134458a190647787e775a46a7ce784fbe98c35ff532671dc5ea08"} Dec 09 11:12:03 crc kubenswrapper[4824]: I1209 11:12:03.708792 4824 scope.go:117] "RemoveContainer" containerID="2318feeaaa390f44a1ec650c04bff1e264f3b1cf059a966897e27f0095b216f4" Dec 09 11:12:17 crc kubenswrapper[4824]: E1209 11:12:17.028399 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Dec 09 11:12:19 crc kubenswrapper[4824]: I1209 11:12:19.971895 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-57nsv"] Dec 09 11:12:19 crc kubenswrapper[4824]: I1209 11:12:19.976375 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-57nsv" Dec 09 11:12:20 crc kubenswrapper[4824]: I1209 11:12:20.004597 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-57nsv"] Dec 09 11:12:20 crc kubenswrapper[4824]: I1209 11:12:20.123108 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9tkb\" (UniqueName: \"kubernetes.io/projected/4cd5d224-1a1e-4154-9b44-4c6b0e9ea657-kube-api-access-g9tkb\") pod \"certified-operators-57nsv\" (UID: \"4cd5d224-1a1e-4154-9b44-4c6b0e9ea657\") " pod="openshift-marketplace/certified-operators-57nsv" Dec 09 11:12:20 crc kubenswrapper[4824]: I1209 11:12:20.123282 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cd5d224-1a1e-4154-9b44-4c6b0e9ea657-utilities\") pod \"certified-operators-57nsv\" (UID: \"4cd5d224-1a1e-4154-9b44-4c6b0e9ea657\") " pod="openshift-marketplace/certified-operators-57nsv" Dec 09 11:12:20 crc kubenswrapper[4824]: I1209 11:12:20.123673 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cd5d224-1a1e-4154-9b44-4c6b0e9ea657-catalog-content\") pod \"certified-operators-57nsv\" (UID: \"4cd5d224-1a1e-4154-9b44-4c6b0e9ea657\") " pod="openshift-marketplace/certified-operators-57nsv" Dec 09 11:12:20 crc kubenswrapper[4824]: I1209 11:12:20.358717 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cd5d224-1a1e-4154-9b44-4c6b0e9ea657-catalog-content\") pod \"certified-operators-57nsv\" (UID: \"4cd5d224-1a1e-4154-9b44-4c6b0e9ea657\") " pod="openshift-marketplace/certified-operators-57nsv" Dec 09 11:12:20 crc kubenswrapper[4824]: I1209 11:12:20.359182 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9tkb\" (UniqueName: \"kubernetes.io/projected/4cd5d224-1a1e-4154-9b44-4c6b0e9ea657-kube-api-access-g9tkb\") pod \"certified-operators-57nsv\" (UID: \"4cd5d224-1a1e-4154-9b44-4c6b0e9ea657\") " pod="openshift-marketplace/certified-operators-57nsv" Dec 09 11:12:20 crc kubenswrapper[4824]: I1209 11:12:20.359293 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cd5d224-1a1e-4154-9b44-4c6b0e9ea657-utilities\") pod \"certified-operators-57nsv\" (UID: \"4cd5d224-1a1e-4154-9b44-4c6b0e9ea657\") " pod="openshift-marketplace/certified-operators-57nsv" Dec 09 11:12:20 crc kubenswrapper[4824]: I1209 11:12:20.359814 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cd5d224-1a1e-4154-9b44-4c6b0e9ea657-utilities\") pod \"certified-operators-57nsv\" (UID: \"4cd5d224-1a1e-4154-9b44-4c6b0e9ea657\") " pod="openshift-marketplace/certified-operators-57nsv" Dec 09 11:12:20 crc kubenswrapper[4824]: I1209 11:12:20.366167 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cd5d224-1a1e-4154-9b44-4c6b0e9ea657-catalog-content\") pod \"certified-operators-57nsv\" (UID: \"4cd5d224-1a1e-4154-9b44-4c6b0e9ea657\") " pod="openshift-marketplace/certified-operators-57nsv" Dec 09 11:12:20 crc kubenswrapper[4824]: I1209 11:12:20.665364 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9tkb\" (UniqueName: \"kubernetes.io/projected/4cd5d224-1a1e-4154-9b44-4c6b0e9ea657-kube-api-access-g9tkb\") pod \"certified-operators-57nsv\" (UID: \"4cd5d224-1a1e-4154-9b44-4c6b0e9ea657\") " pod="openshift-marketplace/certified-operators-57nsv" Dec 09 11:12:20 crc kubenswrapper[4824]: I1209 11:12:20.931760 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-57nsv" Dec 09 11:12:22 crc kubenswrapper[4824]: I1209 11:12:22.289647 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ln79n"] Dec 09 11:12:22 crc kubenswrapper[4824]: I1209 11:12:22.297313 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ln79n" Dec 09 11:12:22 crc kubenswrapper[4824]: I1209 11:12:22.312247 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91c1710e-cd28-4849-b232-9af634a3b317-utilities\") pod \"redhat-marketplace-ln79n\" (UID: \"91c1710e-cd28-4849-b232-9af634a3b317\") " pod="openshift-marketplace/redhat-marketplace-ln79n" Dec 09 11:12:22 crc kubenswrapper[4824]: I1209 11:12:22.312348 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zjn4\" (UniqueName: \"kubernetes.io/projected/91c1710e-cd28-4849-b232-9af634a3b317-kube-api-access-6zjn4\") pod \"redhat-marketplace-ln79n\" (UID: \"91c1710e-cd28-4849-b232-9af634a3b317\") " pod="openshift-marketplace/redhat-marketplace-ln79n" Dec 09 11:12:22 crc kubenswrapper[4824]: I1209 11:12:22.312571 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91c1710e-cd28-4849-b232-9af634a3b317-catalog-content\") pod \"redhat-marketplace-ln79n\" (UID: \"91c1710e-cd28-4849-b232-9af634a3b317\") " pod="openshift-marketplace/redhat-marketplace-ln79n" Dec 09 11:12:22 crc kubenswrapper[4824]: I1209 11:12:22.332048 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ln79n"] Dec 09 11:12:22 crc kubenswrapper[4824]: I1209 11:12:22.468548 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91c1710e-cd28-4849-b232-9af634a3b317-catalog-content\") pod \"redhat-marketplace-ln79n\" (UID: \"91c1710e-cd28-4849-b232-9af634a3b317\") " pod="openshift-marketplace/redhat-marketplace-ln79n" Dec 09 11:12:22 crc kubenswrapper[4824]: I1209 11:12:22.469454 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91c1710e-cd28-4849-b232-9af634a3b317-catalog-content\") pod \"redhat-marketplace-ln79n\" (UID: \"91c1710e-cd28-4849-b232-9af634a3b317\") " pod="openshift-marketplace/redhat-marketplace-ln79n" Dec 09 11:12:22 crc kubenswrapper[4824]: I1209 11:12:22.469576 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91c1710e-cd28-4849-b232-9af634a3b317-utilities\") pod \"redhat-marketplace-ln79n\" (UID: \"91c1710e-cd28-4849-b232-9af634a3b317\") " pod="openshift-marketplace/redhat-marketplace-ln79n" Dec 09 11:12:22 crc kubenswrapper[4824]: I1209 11:12:22.469749 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zjn4\" (UniqueName: \"kubernetes.io/projected/91c1710e-cd28-4849-b232-9af634a3b317-kube-api-access-6zjn4\") pod \"redhat-marketplace-ln79n\" (UID: \"91c1710e-cd28-4849-b232-9af634a3b317\") " pod="openshift-marketplace/redhat-marketplace-ln79n" Dec 09 11:12:22 crc kubenswrapper[4824]: I1209 11:12:22.470051 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91c1710e-cd28-4849-b232-9af634a3b317-utilities\") pod \"redhat-marketplace-ln79n\" (UID: \"91c1710e-cd28-4849-b232-9af634a3b317\") " pod="openshift-marketplace/redhat-marketplace-ln79n" Dec 09 11:12:22 crc kubenswrapper[4824]: I1209 11:12:22.500647 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zjn4\" (UniqueName: \"kubernetes.io/projected/91c1710e-cd28-4849-b232-9af634a3b317-kube-api-access-6zjn4\") pod \"redhat-marketplace-ln79n\" (UID: \"91c1710e-cd28-4849-b232-9af634a3b317\") " pod="openshift-marketplace/redhat-marketplace-ln79n" Dec 09 11:12:22 crc kubenswrapper[4824]: I1209 11:12:22.637445 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ln79n" Dec 09 11:12:33 crc kubenswrapper[4824]: E1209 11:12:33.568496 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 09 11:12:33 crc kubenswrapper[4824]: E1209 11:12:33.569919 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dzfmv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(0fd5955c-4039-42a1-8ecd-0256bb683e61): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 11:12:33 crc kubenswrapper[4824]: E1209 11:12:33.571501 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="0fd5955c-4039-42a1-8ecd-0256bb683e61" Dec 09 11:12:34 crc kubenswrapper[4824]: W1209 11:12:34.197661 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91c1710e_cd28_4849_b232_9af634a3b317.slice/crio-ebacae9b5e0621f05095199a847b87320f2309087a6cfb6bff65b1caae697591 WatchSource:0}: Error finding container ebacae9b5e0621f05095199a847b87320f2309087a6cfb6bff65b1caae697591: Status 404 returned error can't find the container with id ebacae9b5e0621f05095199a847b87320f2309087a6cfb6bff65b1caae697591 Dec 09 11:12:34 crc kubenswrapper[4824]: I1209 11:12:34.209183 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ln79n"] Dec 09 11:12:34 crc kubenswrapper[4824]: W1209 11:12:34.303332 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4cd5d224_1a1e_4154_9b44_4c6b0e9ea657.slice/crio-0d5ba3e0e7f248d6c6c52c35b0c28c4dd8df9da197b8db802e49c647f0bd4ca2 WatchSource:0}: Error finding container 0d5ba3e0e7f248d6c6c52c35b0c28c4dd8df9da197b8db802e49c647f0bd4ca2: Status 404 returned error can't find the container with id 0d5ba3e0e7f248d6c6c52c35b0c28c4dd8df9da197b8db802e49c647f0bd4ca2 Dec 09 11:12:34 crc kubenswrapper[4824]: I1209 11:12:34.304274 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-57nsv"] Dec 09 11:12:34 crc kubenswrapper[4824]: I1209 11:12:34.358931 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ln79n" event={"ID":"91c1710e-cd28-4849-b232-9af634a3b317","Type":"ContainerStarted","Data":"ebacae9b5e0621f05095199a847b87320f2309087a6cfb6bff65b1caae697591"} Dec 09 11:12:34 crc kubenswrapper[4824]: I1209 11:12:34.361534 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerStarted","Data":"84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20"} Dec 09 11:12:34 crc kubenswrapper[4824]: I1209 11:12:34.363539 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57nsv" event={"ID":"4cd5d224-1a1e-4154-9b44-4c6b0e9ea657","Type":"ContainerStarted","Data":"0d5ba3e0e7f248d6c6c52c35b0c28c4dd8df9da197b8db802e49c647f0bd4ca2"} Dec 09 11:12:34 crc kubenswrapper[4824]: E1209 11:12:34.365244 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="0fd5955c-4039-42a1-8ecd-0256bb683e61" Dec 09 11:12:35 crc kubenswrapper[4824]: I1209 11:12:35.380750 4824 generic.go:334] "Generic (PLEG): container finished" podID="91c1710e-cd28-4849-b232-9af634a3b317" containerID="7b42d5048655d1122bddc5990410d23851ce1d1e1d87cf947658c246185d3be7" exitCode=0 Dec 09 11:12:35 crc kubenswrapper[4824]: I1209 11:12:35.381256 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ln79n" event={"ID":"91c1710e-cd28-4849-b232-9af634a3b317","Type":"ContainerDied","Data":"7b42d5048655d1122bddc5990410d23851ce1d1e1d87cf947658c246185d3be7"} Dec 09 11:12:35 crc kubenswrapper[4824]: I1209 11:12:35.384242 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 11:12:35 crc kubenswrapper[4824]: I1209 11:12:35.386695 4824 generic.go:334] "Generic (PLEG): container finished" podID="4cd5d224-1a1e-4154-9b44-4c6b0e9ea657" containerID="422c3df94f97135b56f60816f7af84504df392479759f16b0643e6387d05834e" exitCode=0 Dec 09 11:12:35 crc kubenswrapper[4824]: I1209 11:12:35.388206 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57nsv" event={"ID":"4cd5d224-1a1e-4154-9b44-4c6b0e9ea657","Type":"ContainerDied","Data":"422c3df94f97135b56f60816f7af84504df392479759f16b0643e6387d05834e"} Dec 09 11:12:37 crc kubenswrapper[4824]: I1209 11:12:37.411215 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57nsv" event={"ID":"4cd5d224-1a1e-4154-9b44-4c6b0e9ea657","Type":"ContainerStarted","Data":"1c87327fd72cc9126463a861842f461f38f4cee7cedce8ded539376a6b50984d"} Dec 09 11:12:37 crc kubenswrapper[4824]: I1209 11:12:37.416324 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ln79n" event={"ID":"91c1710e-cd28-4849-b232-9af634a3b317","Type":"ContainerStarted","Data":"93d4f8da0c3eedb8d3da790af410f6df5bcae6d2a5a5bb66d0706b34f54923af"} Dec 09 11:12:39 crc kubenswrapper[4824]: I1209 11:12:39.627464 4824 generic.go:334] "Generic (PLEG): container finished" podID="91c1710e-cd28-4849-b232-9af634a3b317" containerID="93d4f8da0c3eedb8d3da790af410f6df5bcae6d2a5a5bb66d0706b34f54923af" exitCode=0 Dec 09 11:12:39 crc kubenswrapper[4824]: I1209 11:12:39.627667 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ln79n" event={"ID":"91c1710e-cd28-4849-b232-9af634a3b317","Type":"ContainerDied","Data":"93d4f8da0c3eedb8d3da790af410f6df5bcae6d2a5a5bb66d0706b34f54923af"} Dec 09 11:12:40 crc kubenswrapper[4824]: I1209 11:12:40.642247 4824 generic.go:334] "Generic (PLEG): container finished" podID="4cd5d224-1a1e-4154-9b44-4c6b0e9ea657" containerID="1c87327fd72cc9126463a861842f461f38f4cee7cedce8ded539376a6b50984d" exitCode=0 Dec 09 11:12:40 crc kubenswrapper[4824]: I1209 11:12:40.642458 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57nsv" event={"ID":"4cd5d224-1a1e-4154-9b44-4c6b0e9ea657","Type":"ContainerDied","Data":"1c87327fd72cc9126463a861842f461f38f4cee7cedce8ded539376a6b50984d"} Dec 09 11:12:41 crc kubenswrapper[4824]: I1209 11:12:41.656568 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ln79n" event={"ID":"91c1710e-cd28-4849-b232-9af634a3b317","Type":"ContainerStarted","Data":"7d385ec2a9bb49341ac845f835560adc6bfb3fbdfad9268ee4d37ab651f18cce"} Dec 09 11:12:41 crc kubenswrapper[4824]: I1209 11:12:41.659854 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57nsv" event={"ID":"4cd5d224-1a1e-4154-9b44-4c6b0e9ea657","Type":"ContainerStarted","Data":"e66b56a7befee9798e9d59364f1730d449e439faa30a844f807edd1fb1a01a10"} Dec 09 11:12:41 crc kubenswrapper[4824]: I1209 11:12:41.684236 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ln79n" podStartSLOduration=14.449037555 podStartE2EDuration="19.684199984s" podCreationTimestamp="2025-12-09 11:12:22 +0000 UTC" firstStartedPulling="2025-12-09 11:12:35.383853575 +0000 UTC m=+5111.718358242" lastFinishedPulling="2025-12-09 11:12:40.619016004 +0000 UTC m=+5116.953520671" observedRunningTime="2025-12-09 11:12:41.675828311 +0000 UTC m=+5118.010332978" watchObservedRunningTime="2025-12-09 11:12:41.684199984 +0000 UTC m=+5118.018704651" Dec 09 11:12:41 crc kubenswrapper[4824]: I1209 11:12:41.700451 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-57nsv" podStartSLOduration=17.000424722 podStartE2EDuration="22.700424382s" podCreationTimestamp="2025-12-09 11:12:19 +0000 UTC" firstStartedPulling="2025-12-09 11:12:35.389665357 +0000 UTC m=+5111.724170024" lastFinishedPulling="2025-12-09 11:12:41.089665017 +0000 UTC m=+5117.424169684" observedRunningTime="2025-12-09 11:12:41.692418811 +0000 UTC m=+5118.026923498" watchObservedRunningTime="2025-12-09 11:12:41.700424382 +0000 UTC m=+5118.034929049" Dec 09 11:12:42 crc kubenswrapper[4824]: I1209 11:12:42.638061 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ln79n" Dec 09 11:12:42 crc kubenswrapper[4824]: I1209 11:12:42.638309 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ln79n" Dec 09 11:12:43 crc kubenswrapper[4824]: I1209 11:12:43.709241 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-ln79n" podUID="91c1710e-cd28-4849-b232-9af634a3b317" containerName="registry-server" probeResult="failure" output=< Dec 09 11:12:43 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 11:12:43 crc kubenswrapper[4824]: > Dec 09 11:12:48 crc kubenswrapper[4824]: I1209 11:12:48.420521 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 09 11:12:50 crc kubenswrapper[4824]: I1209 11:12:50.932640 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-57nsv" Dec 09 11:12:50 crc kubenswrapper[4824]: I1209 11:12:50.933230 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-57nsv" Dec 09 11:12:51 crc kubenswrapper[4824]: I1209 11:12:51.997168 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-57nsv" podUID="4cd5d224-1a1e-4154-9b44-4c6b0e9ea657" containerName="registry-server" probeResult="failure" output=< Dec 09 11:12:51 crc kubenswrapper[4824]: timeout: health rpc did not complete within 1s Dec 09 11:12:51 crc kubenswrapper[4824]: > Dec 09 11:12:52 crc kubenswrapper[4824]: I1209 11:12:52.681799 4824 patch_prober.go:28] interesting pod/metrics-server-65846dd9c4-g8tlp container/metrics-server namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.79:10250/livez\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:12:52 crc kubenswrapper[4824]: I1209 11:12:52.682281 4824 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-j4zxt container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:12:52 crc kubenswrapper[4824]: I1209 11:12:52.682275 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" podUID="e153e30f-4f5e-4a38-8efb-6452096e25d7" containerName="metrics-server" probeResult="failure" output="Get \"https://10.217.0.79:10250/livez\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:12:52 crc kubenswrapper[4824]: I1209 11:12:52.682343 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" podUID="7b26f121-04f6-4501-84b2-1833b927aa14" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:12:52 crc kubenswrapper[4824]: I1209 11:12:52.697439 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ln79n" Dec 09 11:12:53 crc kubenswrapper[4824]: I1209 11:12:53.056049 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ln79n" Dec 09 11:12:53 crc kubenswrapper[4824]: I1209 11:12:53.194961 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"0fd5955c-4039-42a1-8ecd-0256bb683e61","Type":"ContainerStarted","Data":"a8a574116ea94300bb3b084f452cedabd53a1b7b6e69d43cda4dd982221f00fe"} Dec 09 11:12:53 crc kubenswrapper[4824]: I1209 11:12:53.488634 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ln79n"] Dec 09 11:12:54 crc kubenswrapper[4824]: I1209 11:12:54.203719 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ln79n" podUID="91c1710e-cd28-4849-b232-9af634a3b317" containerName="registry-server" containerID="cri-o://7d385ec2a9bb49341ac845f835560adc6bfb3fbdfad9268ee4d37ab651f18cce" gracePeriod=2 Dec 09 11:12:54 crc kubenswrapper[4824]: I1209 11:12:54.236274 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=8.032340229 podStartE2EDuration="1m12.236253846s" podCreationTimestamp="2025-12-09 11:11:42 +0000 UTC" firstStartedPulling="2025-12-09 11:11:44.21304683 +0000 UTC m=+5060.547551497" lastFinishedPulling="2025-12-09 11:12:48.416960447 +0000 UTC m=+5124.751465114" observedRunningTime="2025-12-09 11:12:54.220873645 +0000 UTC m=+5130.555378322" watchObservedRunningTime="2025-12-09 11:12:54.236253846 +0000 UTC m=+5130.570758513" Dec 09 11:12:54 crc kubenswrapper[4824]: I1209 11:12:54.894838 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ln79n" Dec 09 11:12:54 crc kubenswrapper[4824]: I1209 11:12:54.976047 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91c1710e-cd28-4849-b232-9af634a3b317-utilities\") pod \"91c1710e-cd28-4849-b232-9af634a3b317\" (UID: \"91c1710e-cd28-4849-b232-9af634a3b317\") " Dec 09 11:12:54 crc kubenswrapper[4824]: I1209 11:12:54.976221 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zjn4\" (UniqueName: \"kubernetes.io/projected/91c1710e-cd28-4849-b232-9af634a3b317-kube-api-access-6zjn4\") pod \"91c1710e-cd28-4849-b232-9af634a3b317\" (UID: \"91c1710e-cd28-4849-b232-9af634a3b317\") " Dec 09 11:12:54 crc kubenswrapper[4824]: I1209 11:12:54.976370 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91c1710e-cd28-4849-b232-9af634a3b317-catalog-content\") pod \"91c1710e-cd28-4849-b232-9af634a3b317\" (UID: \"91c1710e-cd28-4849-b232-9af634a3b317\") " Dec 09 11:12:54 crc kubenswrapper[4824]: I1209 11:12:54.976753 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91c1710e-cd28-4849-b232-9af634a3b317-utilities" (OuterVolumeSpecName: "utilities") pod "91c1710e-cd28-4849-b232-9af634a3b317" (UID: "91c1710e-cd28-4849-b232-9af634a3b317"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:12:54 crc kubenswrapper[4824]: I1209 11:12:54.979022 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91c1710e-cd28-4849-b232-9af634a3b317-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:12:54 crc kubenswrapper[4824]: I1209 11:12:54.983274 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91c1710e-cd28-4849-b232-9af634a3b317-kube-api-access-6zjn4" (OuterVolumeSpecName: "kube-api-access-6zjn4") pod "91c1710e-cd28-4849-b232-9af634a3b317" (UID: "91c1710e-cd28-4849-b232-9af634a3b317"). InnerVolumeSpecName "kube-api-access-6zjn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:12:55 crc kubenswrapper[4824]: I1209 11:12:55.028250 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91c1710e-cd28-4849-b232-9af634a3b317-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91c1710e-cd28-4849-b232-9af634a3b317" (UID: "91c1710e-cd28-4849-b232-9af634a3b317"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:12:55 crc kubenswrapper[4824]: I1209 11:12:55.081805 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91c1710e-cd28-4849-b232-9af634a3b317-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:12:55 crc kubenswrapper[4824]: I1209 11:12:55.081841 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zjn4\" (UniqueName: \"kubernetes.io/projected/91c1710e-cd28-4849-b232-9af634a3b317-kube-api-access-6zjn4\") on node \"crc\" DevicePath \"\"" Dec 09 11:12:55 crc kubenswrapper[4824]: I1209 11:12:55.220882 4824 generic.go:334] "Generic (PLEG): container finished" podID="91c1710e-cd28-4849-b232-9af634a3b317" containerID="7d385ec2a9bb49341ac845f835560adc6bfb3fbdfad9268ee4d37ab651f18cce" exitCode=0 Dec 09 11:12:55 crc kubenswrapper[4824]: I1209 11:12:55.220941 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ln79n" Dec 09 11:12:55 crc kubenswrapper[4824]: I1209 11:12:55.220941 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ln79n" event={"ID":"91c1710e-cd28-4849-b232-9af634a3b317","Type":"ContainerDied","Data":"7d385ec2a9bb49341ac845f835560adc6bfb3fbdfad9268ee4d37ab651f18cce"} Dec 09 11:12:55 crc kubenswrapper[4824]: I1209 11:12:55.220988 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ln79n" event={"ID":"91c1710e-cd28-4849-b232-9af634a3b317","Type":"ContainerDied","Data":"ebacae9b5e0621f05095199a847b87320f2309087a6cfb6bff65b1caae697591"} Dec 09 11:12:55 crc kubenswrapper[4824]: I1209 11:12:55.221022 4824 scope.go:117] "RemoveContainer" containerID="7d385ec2a9bb49341ac845f835560adc6bfb3fbdfad9268ee4d37ab651f18cce" Dec 09 11:12:55 crc kubenswrapper[4824]: I1209 11:12:55.698157 4824 scope.go:117] "RemoveContainer" containerID="93d4f8da0c3eedb8d3da790af410f6df5bcae6d2a5a5bb66d0706b34f54923af" Dec 09 11:12:55 crc kubenswrapper[4824]: I1209 11:12:55.709830 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ln79n"] Dec 09 11:12:55 crc kubenswrapper[4824]: I1209 11:12:55.743303 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ln79n"] Dec 09 11:12:55 crc kubenswrapper[4824]: I1209 11:12:55.765402 4824 scope.go:117] "RemoveContainer" containerID="7b42d5048655d1122bddc5990410d23851ce1d1e1d87cf947658c246185d3be7" Dec 09 11:12:55 crc kubenswrapper[4824]: I1209 11:12:55.831020 4824 scope.go:117] "RemoveContainer" containerID="7d385ec2a9bb49341ac845f835560adc6bfb3fbdfad9268ee4d37ab651f18cce" Dec 09 11:12:55 crc kubenswrapper[4824]: E1209 11:12:55.831539 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d385ec2a9bb49341ac845f835560adc6bfb3fbdfad9268ee4d37ab651f18cce\": container with ID starting with 7d385ec2a9bb49341ac845f835560adc6bfb3fbdfad9268ee4d37ab651f18cce not found: ID does not exist" containerID="7d385ec2a9bb49341ac845f835560adc6bfb3fbdfad9268ee4d37ab651f18cce" Dec 09 11:12:55 crc kubenswrapper[4824]: I1209 11:12:55.831585 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d385ec2a9bb49341ac845f835560adc6bfb3fbdfad9268ee4d37ab651f18cce"} err="failed to get container status \"7d385ec2a9bb49341ac845f835560adc6bfb3fbdfad9268ee4d37ab651f18cce\": rpc error: code = NotFound desc = could not find container \"7d385ec2a9bb49341ac845f835560adc6bfb3fbdfad9268ee4d37ab651f18cce\": container with ID starting with 7d385ec2a9bb49341ac845f835560adc6bfb3fbdfad9268ee4d37ab651f18cce not found: ID does not exist" Dec 09 11:12:55 crc kubenswrapper[4824]: I1209 11:12:55.831622 4824 scope.go:117] "RemoveContainer" containerID="93d4f8da0c3eedb8d3da790af410f6df5bcae6d2a5a5bb66d0706b34f54923af" Dec 09 11:12:55 crc kubenswrapper[4824]: E1209 11:12:55.831898 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93d4f8da0c3eedb8d3da790af410f6df5bcae6d2a5a5bb66d0706b34f54923af\": container with ID starting with 93d4f8da0c3eedb8d3da790af410f6df5bcae6d2a5a5bb66d0706b34f54923af not found: ID does not exist" containerID="93d4f8da0c3eedb8d3da790af410f6df5bcae6d2a5a5bb66d0706b34f54923af" Dec 09 11:12:55 crc kubenswrapper[4824]: I1209 11:12:55.831924 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93d4f8da0c3eedb8d3da790af410f6df5bcae6d2a5a5bb66d0706b34f54923af"} err="failed to get container status \"93d4f8da0c3eedb8d3da790af410f6df5bcae6d2a5a5bb66d0706b34f54923af\": rpc error: code = NotFound desc = could not find container \"93d4f8da0c3eedb8d3da790af410f6df5bcae6d2a5a5bb66d0706b34f54923af\": container with ID starting with 93d4f8da0c3eedb8d3da790af410f6df5bcae6d2a5a5bb66d0706b34f54923af not found: ID does not exist" Dec 09 11:12:55 crc kubenswrapper[4824]: I1209 11:12:55.831938 4824 scope.go:117] "RemoveContainer" containerID="7b42d5048655d1122bddc5990410d23851ce1d1e1d87cf947658c246185d3be7" Dec 09 11:12:55 crc kubenswrapper[4824]: E1209 11:12:55.832229 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b42d5048655d1122bddc5990410d23851ce1d1e1d87cf947658c246185d3be7\": container with ID starting with 7b42d5048655d1122bddc5990410d23851ce1d1e1d87cf947658c246185d3be7 not found: ID does not exist" containerID="7b42d5048655d1122bddc5990410d23851ce1d1e1d87cf947658c246185d3be7" Dec 09 11:12:55 crc kubenswrapper[4824]: I1209 11:12:55.832258 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b42d5048655d1122bddc5990410d23851ce1d1e1d87cf947658c246185d3be7"} err="failed to get container status \"7b42d5048655d1122bddc5990410d23851ce1d1e1d87cf947658c246185d3be7\": rpc error: code = NotFound desc = could not find container \"7b42d5048655d1122bddc5990410d23851ce1d1e1d87cf947658c246185d3be7\": container with ID starting with 7b42d5048655d1122bddc5990410d23851ce1d1e1d87cf947658c246185d3be7 not found: ID does not exist" Dec 09 11:12:55 crc kubenswrapper[4824]: I1209 11:12:55.941352 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91c1710e-cd28-4849-b232-9af634a3b317" path="/var/lib/kubelet/pods/91c1710e-cd28-4849-b232-9af634a3b317/volumes" Dec 09 11:13:01 crc kubenswrapper[4824]: I1209 11:13:01.035731 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-57nsv" Dec 09 11:13:01 crc kubenswrapper[4824]: I1209 11:13:01.115396 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-57nsv" Dec 09 11:13:01 crc kubenswrapper[4824]: I1209 11:13:01.299092 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-57nsv"] Dec 09 11:13:02 crc kubenswrapper[4824]: I1209 11:13:02.297176 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-57nsv" podUID="4cd5d224-1a1e-4154-9b44-4c6b0e9ea657" containerName="registry-server" containerID="cri-o://e66b56a7befee9798e9d59364f1730d449e439faa30a844f807edd1fb1a01a10" gracePeriod=2 Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.221874 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-57nsv" Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.339670 4824 generic.go:334] "Generic (PLEG): container finished" podID="4cd5d224-1a1e-4154-9b44-4c6b0e9ea657" containerID="e66b56a7befee9798e9d59364f1730d449e439faa30a844f807edd1fb1a01a10" exitCode=0 Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.339741 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57nsv" event={"ID":"4cd5d224-1a1e-4154-9b44-4c6b0e9ea657","Type":"ContainerDied","Data":"e66b56a7befee9798e9d59364f1730d449e439faa30a844f807edd1fb1a01a10"} Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.339772 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57nsv" event={"ID":"4cd5d224-1a1e-4154-9b44-4c6b0e9ea657","Type":"ContainerDied","Data":"0d5ba3e0e7f248d6c6c52c35b0c28c4dd8df9da197b8db802e49c647f0bd4ca2"} Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.339807 4824 scope.go:117] "RemoveContainer" containerID="e66b56a7befee9798e9d59364f1730d449e439faa30a844f807edd1fb1a01a10" Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.339888 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-57nsv" Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.367584 4824 scope.go:117] "RemoveContainer" containerID="1c87327fd72cc9126463a861842f461f38f4cee7cedce8ded539376a6b50984d" Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.398919 4824 scope.go:117] "RemoveContainer" containerID="422c3df94f97135b56f60816f7af84504df392479759f16b0643e6387d05834e" Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.472447 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9tkb\" (UniqueName: \"kubernetes.io/projected/4cd5d224-1a1e-4154-9b44-4c6b0e9ea657-kube-api-access-g9tkb\") pod \"4cd5d224-1a1e-4154-9b44-4c6b0e9ea657\" (UID: \"4cd5d224-1a1e-4154-9b44-4c6b0e9ea657\") " Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.473459 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cd5d224-1a1e-4154-9b44-4c6b0e9ea657-utilities\") pod \"4cd5d224-1a1e-4154-9b44-4c6b0e9ea657\" (UID: \"4cd5d224-1a1e-4154-9b44-4c6b0e9ea657\") " Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.473552 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cd5d224-1a1e-4154-9b44-4c6b0e9ea657-catalog-content\") pod \"4cd5d224-1a1e-4154-9b44-4c6b0e9ea657\" (UID: \"4cd5d224-1a1e-4154-9b44-4c6b0e9ea657\") " Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.474801 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cd5d224-1a1e-4154-9b44-4c6b0e9ea657-utilities" (OuterVolumeSpecName: "utilities") pod "4cd5d224-1a1e-4154-9b44-4c6b0e9ea657" (UID: "4cd5d224-1a1e-4154-9b44-4c6b0e9ea657"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.480989 4824 scope.go:117] "RemoveContainer" containerID="e66b56a7befee9798e9d59364f1730d449e439faa30a844f807edd1fb1a01a10" Dec 09 11:13:03 crc kubenswrapper[4824]: E1209 11:13:03.482448 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e66b56a7befee9798e9d59364f1730d449e439faa30a844f807edd1fb1a01a10\": container with ID starting with e66b56a7befee9798e9d59364f1730d449e439faa30a844f807edd1fb1a01a10 not found: ID does not exist" containerID="e66b56a7befee9798e9d59364f1730d449e439faa30a844f807edd1fb1a01a10" Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.482484 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e66b56a7befee9798e9d59364f1730d449e439faa30a844f807edd1fb1a01a10"} err="failed to get container status \"e66b56a7befee9798e9d59364f1730d449e439faa30a844f807edd1fb1a01a10\": rpc error: code = NotFound desc = could not find container \"e66b56a7befee9798e9d59364f1730d449e439faa30a844f807edd1fb1a01a10\": container with ID starting with e66b56a7befee9798e9d59364f1730d449e439faa30a844f807edd1fb1a01a10 not found: ID does not exist" Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.482507 4824 scope.go:117] "RemoveContainer" containerID="1c87327fd72cc9126463a861842f461f38f4cee7cedce8ded539376a6b50984d" Dec 09 11:13:03 crc kubenswrapper[4824]: E1209 11:13:03.484903 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c87327fd72cc9126463a861842f461f38f4cee7cedce8ded539376a6b50984d\": container with ID starting with 1c87327fd72cc9126463a861842f461f38f4cee7cedce8ded539376a6b50984d not found: ID does not exist" containerID="1c87327fd72cc9126463a861842f461f38f4cee7cedce8ded539376a6b50984d" Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.484941 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c87327fd72cc9126463a861842f461f38f4cee7cedce8ded539376a6b50984d"} err="failed to get container status \"1c87327fd72cc9126463a861842f461f38f4cee7cedce8ded539376a6b50984d\": rpc error: code = NotFound desc = could not find container \"1c87327fd72cc9126463a861842f461f38f4cee7cedce8ded539376a6b50984d\": container with ID starting with 1c87327fd72cc9126463a861842f461f38f4cee7cedce8ded539376a6b50984d not found: ID does not exist" Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.484962 4824 scope.go:117] "RemoveContainer" containerID="422c3df94f97135b56f60816f7af84504df392479759f16b0643e6387d05834e" Dec 09 11:13:03 crc kubenswrapper[4824]: E1209 11:13:03.488959 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"422c3df94f97135b56f60816f7af84504df392479759f16b0643e6387d05834e\": container with ID starting with 422c3df94f97135b56f60816f7af84504df392479759f16b0643e6387d05834e not found: ID does not exist" containerID="422c3df94f97135b56f60816f7af84504df392479759f16b0643e6387d05834e" Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.489016 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"422c3df94f97135b56f60816f7af84504df392479759f16b0643e6387d05834e"} err="failed to get container status \"422c3df94f97135b56f60816f7af84504df392479759f16b0643e6387d05834e\": rpc error: code = NotFound desc = could not find container \"422c3df94f97135b56f60816f7af84504df392479759f16b0643e6387d05834e\": container with ID starting with 422c3df94f97135b56f60816f7af84504df392479759f16b0643e6387d05834e not found: ID does not exist" Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.497796 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cd5d224-1a1e-4154-9b44-4c6b0e9ea657-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.519040 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cd5d224-1a1e-4154-9b44-4c6b0e9ea657-kube-api-access-g9tkb" (OuterVolumeSpecName: "kube-api-access-g9tkb") pod "4cd5d224-1a1e-4154-9b44-4c6b0e9ea657" (UID: "4cd5d224-1a1e-4154-9b44-4c6b0e9ea657"). InnerVolumeSpecName "kube-api-access-g9tkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.599647 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9tkb\" (UniqueName: \"kubernetes.io/projected/4cd5d224-1a1e-4154-9b44-4c6b0e9ea657-kube-api-access-g9tkb\") on node \"crc\" DevicePath \"\"" Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.665553 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cd5d224-1a1e-4154-9b44-4c6b0e9ea657-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4cd5d224-1a1e-4154-9b44-4c6b0e9ea657" (UID: "4cd5d224-1a1e-4154-9b44-4c6b0e9ea657"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.702201 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cd5d224-1a1e-4154-9b44-4c6b0e9ea657-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.969646 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-57nsv"] Dec 09 11:13:03 crc kubenswrapper[4824]: I1209 11:13:03.980243 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-57nsv"] Dec 09 11:13:05 crc kubenswrapper[4824]: I1209 11:13:05.928964 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cd5d224-1a1e-4154-9b44-4c6b0e9ea657" path="/var/lib/kubelet/pods/4cd5d224-1a1e-4154-9b44-4c6b0e9ea657/volumes" Dec 09 11:13:48 crc kubenswrapper[4824]: I1209 11:13:48.747497 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="56148e0f-636f-410e-bfb5-342da01e8c76" containerName="galera" probeResult="failure" output="command timed out" Dec 09 11:13:48 crc kubenswrapper[4824]: I1209 11:13:48.747993 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="56148e0f-636f-410e-bfb5-342da01e8c76" containerName="galera" probeResult="failure" output="command timed out" Dec 09 11:13:52 crc kubenswrapper[4824]: I1209 11:13:52.046813 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6mfg8" podUID="5b34da01-28d8-476c-a2a5-a098489d90c1" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.102:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:52 crc kubenswrapper[4824]: I1209 11:13:52.046836 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4" podUID="143e48cd-c956-448c-8dcb-5858a582e29c" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.104:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:52 crc kubenswrapper[4824]: I1209 11:13:52.048529 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6mfg8" podUID="5b34da01-28d8-476c-a2a5-a098489d90c1" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.102:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:52 crc kubenswrapper[4824]: I1209 11:13:52.049673 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4" podUID="143e48cd-c956-448c-8dcb-5858a582e29c" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.104:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:52 crc kubenswrapper[4824]: I1209 11:13:52.287086 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-967d97867-lfprd" podUID="b0c20e13-b233-4b21-8f47-facef9603735" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:52 crc kubenswrapper[4824]: I1209 11:13:52.287301 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/ironic-operator-controller-manager-967d97867-lfprd" podUID="b0c20e13-b233-4b21-8f47-facef9603735" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:52 crc kubenswrapper[4824]: I1209 11:13:52.287505 4824 patch_prober.go:28] interesting pod/nmstate-webhook-5f6d4c5ccb-jklpr container/nmstate-webhook namespace/openshift-nmstate: Readiness probe status=failure output="Get \"https://10.217.0.88:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:52 crc kubenswrapper[4824]: I1209 11:13:52.289049 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jklpr" podUID="0b932274-5eba-4e0d-8a64-1f469a6ab3d1" containerName="nmstate-webhook" probeResult="failure" output="Get \"https://10.217.0.88:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:52 crc kubenswrapper[4824]: I1209 11:13:52.499372 4824 patch_prober.go:28] interesting pod/console-operator-58897d9998-wgtx2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:52 crc kubenswrapper[4824]: I1209 11:13:52.499453 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" podUID="cd6c14a0-99d8-485d-81f3-cc4ebe5a943b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:52 crc kubenswrapper[4824]: I1209 11:13:52.499486 4824 patch_prober.go:28] interesting pod/console-operator-58897d9998-wgtx2 container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:52 crc kubenswrapper[4824]: I1209 11:13:52.499587 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" podUID="cd6c14a0-99d8-485d-81f3-cc4ebe5a943b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:52 crc kubenswrapper[4824]: I1209 11:13:52.581989 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl" podUID="b6577050-6a21-43bb-84aa-20aad247aafc" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:52 crc kubenswrapper[4824]: I1209 11:13:52.582055 4824 patch_prober.go:28] interesting pod/metrics-server-65846dd9c4-g8tlp container/metrics-server namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.79:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:52 crc kubenswrapper[4824]: I1209 11:13:52.582171 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" podUID="e153e30f-4f5e-4a38-8efb-6452096e25d7" containerName="metrics-server" probeResult="failure" output="Get \"https://10.217.0.79:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:52 crc kubenswrapper[4824]: I1209 11:13:52.582208 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl" podUID="b6577050-6a21-43bb-84aa-20aad247aafc" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:52 crc kubenswrapper[4824]: I1209 11:13:52.807307 4824 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-bxplc container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:52 crc kubenswrapper[4824]: I1209 11:13:52.807637 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" podUID="c3400b7c-5666-4b36-a50f-7800f6737527" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:52 crc kubenswrapper[4824]: I1209 11:13:52.913035 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8" podUID="23621a98-1d6b-421a-976c-965cecfe6db1" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.113:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:52.995965 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8" podUID="23621a98-1d6b-421a-976c-965cecfe6db1" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.113:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:52.996004 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" podUID="d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:53.134141 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" podUID="383a67da-0f83-41e2-82d6-809d1b6cdbec" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.118:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:53.134564 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" podUID="d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:53.301115 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl" podUID="6e91c44f-6221-4dfd-9c18-2bbedff02850" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.119:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:53.301620 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh" podUID="91bf6cd9-a4fe-46e8-b1e7-39e57c1398b3" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:53.301872 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh" podUID="91bf6cd9-a4fe-46e8-b1e7-39e57c1398b3" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:53.301895 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" podUID="383a67da-0f83-41e2-82d6-809d1b6cdbec" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.118:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:53.431953 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" podUID="bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.121:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:53.472976 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" podUID="feab57e6-788b-4f23-9e99-aa248786052d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:53.473100 4824 patch_prober.go:28] interesting pod/monitoring-plugin-6f74c7bb76-pbw94 container/monitoring-plugin namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.80:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:53.473154 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/monitoring-plugin-6f74c7bb76-pbw94" podUID="364b1858-6a46-4db4-9d9d-0ec19d54abc9" containerName="monitoring-plugin" probeResult="failure" output="Get \"https://10.217.0.80:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:53.473207 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9" podUID="31749799-9eb4-403f-a61d-9d50d8bc8367" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.117:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:53.473436 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl" podUID="6e91c44f-6221-4dfd-9c18-2bbedff02850" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.119:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:53.555428 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" podUID="9f87669d-33a9-4269-808c-ca6c718f762c" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.44:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:53.556231 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9" podUID="31749799-9eb4-403f-a61d-9d50d8bc8367" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.117:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:53.639111 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-5xfv2" podUID="87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:53.639325 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" podUID="bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.121:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:53.639414 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" podUID="feab57e6-788b-4f23-9e99-aa248786052d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:53.640311 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" podUID="9f87669d-33a9-4269-808c-ca6c718f762c" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.44:6080/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:53.640340 4824 patch_prober.go:28] interesting pod/logging-loki-distributor-76cc67bf56-nl2vt container/loki-distributor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.51:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:53.640389 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" podUID="72ff7b94-63fd-45d1-a803-bc1ccf4388ed" containerName="loki-distributor" probeResult="failure" output="Get \"https://10.217.0.51:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:53.640436 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-5xfv2" podUID="87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:53.780035 4824 patch_prober.go:28] interesting pod/logging-loki-querier-5895d59bb8-89dgf container/loki-querier namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.52:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:53 crc kubenswrapper[4824]: I1209 11:13:53.780453 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" podUID="1006b7c9-7244-44af-8bc9-52787d891f7f" containerName="loki-querier" probeResult="failure" output="Get \"https://10.217.0.52:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:53.918101 4824 patch_prober.go:28] interesting pod/router-default-5444994796-mj7jx container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:53.918170 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:53.918101 4824 patch_prober.go:28] interesting pod/router-default-5444994796-mj7jx container/router namespace/openshift-ingress: Liveness probe status=failure output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:53.918287 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:53.950166 4824 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-tg5lv container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:53.950223 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" podUID="45732d49-ce14-439b-8b7a-fb9fdf267fb1" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:53.950165 4824 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-tg5lv container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:53.950283 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" podUID="45732d49-ce14-439b-8b7a-fb9fdf267fb1" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.049065 4824 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-9c2mr container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.049135 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" podUID="62823972-71dc-4f7d-b4f6-da0cb90dcf36" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.049201 4824 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-k8bgl container/package-server-manager namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.049218 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" podUID="6c10edf9-cf75-4633-8d4c-9bbb9d9b6339" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.049262 4824 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-9c2mr container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.049275 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" podUID="62823972-71dc-4f7d-b4f6-da0cb90dcf36" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.049309 4824 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-c2dfm container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.049325 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c2dfm" podUID="a7c02c2b-70ae-4b6f-85ae-5a38b737a73b" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.049355 4824 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-c2dfm container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.049369 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c2dfm" podUID="a7c02c2b-70ae-4b6f-85ae-5a38b737a73b" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.049396 4824 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-k8bgl container/package-server-manager namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.049412 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" podUID="6c10edf9-cf75-4633-8d4c-9bbb9d9b6339" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.568531 4824 patch_prober.go:28] interesting pod/logging-loki-query-frontend-84558f7c9f-9d5fm container/loki-query-frontend namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.568625 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" podUID="e05c11bd-cd92-4bac-adea-0a6049ccfb39" containerName="loki-query-frontend" probeResult="failure" output="Get \"https://10.217.0.53:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.690011 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-vz44w container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.690086 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-vz44w container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:8081/ready\": context deadline exceeded" start-of-body= Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.690161 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" podUID="93c989b8-7b86-4339-bbab-5886c7d13dc9" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.55:8081/ready\": context deadline exceeded" Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.690087 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" podUID="93c989b8-7b86-4339-bbab-5886c7d13dc9" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.55:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.717716 4824 patch_prober.go:28] interesting pod/console-7c85c585b4-tgrff container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.139:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.717833 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-7c85c585b4-tgrff" podUID="e61278e7-d0a8-4039-909d-c2812c8a4a81" containerName="console" probeResult="failure" output="Get \"https://10.217.0.139:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.847067 4824 patch_prober.go:28] interesting pod/loki-operator-controller-manager-6bf69c47b7-8fjw7 container/manager namespace/openshift-operators-redhat: Readiness probe status=failure output="Get \"http://10.217.0.47:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.847422 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" podUID="352c7865-5a0b-4ebb-93f2-513ce433ff8f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.47:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.850698 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-bvp6v container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8081/ready\": context deadline exceeded" start-of-body= Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.850735 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" podUID="858d899b-800f-4639-8fbd-4f1ccad44991" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.54:8081/ready\": context deadline exceeded" Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.851383 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-bvp6v container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:54 crc kubenswrapper[4824]: I1209 11:13:54.851459 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" podUID="858d899b-800f-4639-8fbd-4f1ccad44991" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:55 crc kubenswrapper[4824]: I1209 11:13:55.257483 4824 patch_prober.go:28] interesting pod/oauth-openshift-f578d5c8f-5jzzv container/oauth-openshift namespace/openshift-authentication: Liveness probe status=failure output="Get \"https://10.217.0.62:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:55 crc kubenswrapper[4824]: I1209 11:13:55.257552 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" podUID="3d1e2cd1-5fc5-4a76-b1ff-07db11b07708" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.62:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:55 crc kubenswrapper[4824]: I1209 11:13:55.258093 4824 patch_prober.go:28] interesting pod/oauth-openshift-f578d5c8f-5jzzv container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.62:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:55 crc kubenswrapper[4824]: I1209 11:13:55.258184 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" podUID="3d1e2cd1-5fc5-4a76-b1ff-07db11b07708" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.62:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:55 crc kubenswrapper[4824]: I1209 11:13:55.309550 4824 patch_prober.go:28] interesting pod/logging-loki-compactor-0 container/loki-compactor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.65:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:55 crc kubenswrapper[4824]: I1209 11:13:55.309626 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-compactor-0" podUID="379b9136-1cdf-4786-8494-e99c7161b202" containerName="loki-compactor" probeResult="failure" output="Get \"https://10.217.0.65:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:55 crc kubenswrapper[4824]: I1209 11:13:55.374457 4824 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.64:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:55 crc kubenswrapper[4824]: I1209 11:13:55.374838 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="9d24f296-729f-4eec-a1ae-d6b904399394" containerName="loki-ingester" probeResult="failure" output="Get \"https://10.217.0.64:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:55 crc kubenswrapper[4824]: I1209 11:13:55.453879 4824 patch_prober.go:28] interesting pod/logging-loki-index-gateway-0 container/loki-index-gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.74:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:55 crc kubenswrapper[4824]: I1209 11:13:55.453966 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-index-gateway-0" podUID="6baf0bfe-7be6-4309-93da-e86174e4654e" containerName="loki-index-gateway" probeResult="failure" output="Get \"https://10.217.0.74:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:55 crc kubenswrapper[4824]: I1209 11:13:55.908033 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" podUID="e7e496dc-23c4-47cd-9c5a-aa4430130849" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.123:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:56 crc kubenswrapper[4824]: I1209 11:13:56.945171 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-95jfr" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" containerName="registry-server" probeResult="failure" output=< Dec 09 11:13:56 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 11:13:56 crc kubenswrapper[4824]: > Dec 09 11:13:56 crc kubenswrapper[4824]: I1209 11:13:56.945199 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-marketplace-95jfr" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" containerName="registry-server" probeResult="failure" output=< Dec 09 11:13:56 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 11:13:56 crc kubenswrapper[4824]: > Dec 09 11:13:56 crc kubenswrapper[4824]: I1209 11:13:56.945195 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-operators-582xj" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" containerName="registry-server" probeResult="failure" output=< Dec 09 11:13:56 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 11:13:56 crc kubenswrapper[4824]: > Dec 09 11:13:56 crc kubenswrapper[4824]: I1209 11:13:56.945829 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-northd-0" podUID="c2dd0680-d32f-45e8-b834-6a327d2eaa21" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 09 11:13:56 crc kubenswrapper[4824]: I1209 11:13:56.945848 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="c2dd0680-d32f-45e8-b834-6a327d2eaa21" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 09 11:13:57 crc kubenswrapper[4824]: I1209 11:13:57.054946 4824 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:57 crc kubenswrapper[4824]: I1209 11:13:57.055040 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:57 crc kubenswrapper[4824]: I1209 11:13:57.097155 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/kube-state-metrics-0" podUID="54621281-db00-41e4-b617-032435893391" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.0.254:8080/livez\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:57 crc kubenswrapper[4824]: I1209 11:13:57.097350 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="54621281-db00-41e4-b617-032435893391" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.0.254:8081/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:57 crc kubenswrapper[4824]: I1209 11:13:57.112041 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/prometheus-metric-storage-0" podUID="fb8f1e44-4c21-423a-bacc-1cca7d7715c3" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.167:9090/-/healthy\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:57 crc kubenswrapper[4824]: I1209 11:13:57.112041 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="fb8f1e44-4c21-423a-bacc-1cca7d7715c3" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.167:9090/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:57 crc kubenswrapper[4824]: I1209 11:13:57.171838 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-operators-582xj" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" containerName="registry-server" probeResult="failure" output=< Dec 09 11:13:57 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 11:13:57 crc kubenswrapper[4824]: > Dec 09 11:13:57 crc kubenswrapper[4824]: I1209 11:13:57.744290 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-handler-5strj" podUID="952171d0-1a2e-4801-bc11-4012ba19588c" containerName="nmstate-handler" probeResult="failure" output="command timed out" Dec 09 11:13:57 crc kubenswrapper[4824]: I1209 11:13:57.945555 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/community-operators-jm5ld" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" containerName="registry-server" probeResult="failure" output=< Dec 09 11:13:57 crc kubenswrapper[4824]: timeout: health rpc did not complete within 1s Dec 09 11:13:57 crc kubenswrapper[4824]: > Dec 09 11:13:57 crc kubenswrapper[4824]: I1209 11:13:57.946756 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-jm5ld" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" containerName="registry-server" probeResult="failure" output=< Dec 09 11:13:57 crc kubenswrapper[4824]: timeout: health rpc did not complete within 1s Dec 09 11:13:57 crc kubenswrapper[4824]: > Dec 09 11:13:57 crc kubenswrapper[4824]: I1209 11:13:57.947173 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/certified-operators-2cbmd" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" containerName="registry-server" probeResult="failure" output=< Dec 09 11:13:57 crc kubenswrapper[4824]: timeout: health rpc did not complete within 1s Dec 09 11:13:57 crc kubenswrapper[4824]: > Dec 09 11:13:57 crc kubenswrapper[4824]: I1209 11:13:57.948174 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/certified-operators-2cbmd" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" containerName="registry-server" probeResult="failure" output=< Dec 09 11:13:57 crc kubenswrapper[4824]: timeout: health rpc did not complete within 1s Dec 09 11:13:57 crc kubenswrapper[4824]: > Dec 09 11:13:58 crc kubenswrapper[4824]: I1209 11:13:58.021103 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" podUID="1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.108:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:58 crc kubenswrapper[4824]: I1209 11:13:58.300286 4824 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-4dlng container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:58 crc kubenswrapper[4824]: I1209 11:13:58.300361 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" podUID="86c42164-acb8-4a27-9a0f-48307cd304e4" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:58 crc kubenswrapper[4824]: I1209 11:13:58.300286 4824 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-4dlng container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:58 crc kubenswrapper[4824]: I1209 11:13:58.300428 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" podUID="86c42164-acb8-4a27-9a0f-48307cd304e4" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:58 crc kubenswrapper[4824]: I1209 11:13:58.591072 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" podUID="c2889de3-49b2-4465-8235-37f9e58b42a3" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.116:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:58 crc kubenswrapper[4824]: I1209 11:13:58.744392 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="56148e0f-636f-410e-bfb5-342da01e8c76" containerName="galera" probeResult="failure" output="command timed out" Dec 09 11:13:58 crc kubenswrapper[4824]: I1209 11:13:58.745779 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="56148e0f-636f-410e-bfb5-342da01e8c76" containerName="galera" probeResult="failure" output="command timed out" Dec 09 11:13:58 crc kubenswrapper[4824]: I1209 11:13:58.749158 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 09 11:13:59 crc kubenswrapper[4824]: I1209 11:13:59.304913 4824 patch_prober.go:28] interesting pod/route-controller-manager-7f84c5b8f9-rqmr9 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:59 crc kubenswrapper[4824]: I1209 11:13:59.304946 4824 patch_prober.go:28] interesting pod/route-controller-manager-7f84c5b8f9-rqmr9 container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:59 crc kubenswrapper[4824]: I1209 11:13:59.304999 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" podUID="afe92d2b-bce3-445e-9c26-9c533353e7a3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:59 crc kubenswrapper[4824]: I1209 11:13:59.305020 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" podUID="afe92d2b-bce3-445e-9c26-9c533353e7a3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:59 crc kubenswrapper[4824]: I1209 11:13:59.689231 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-vz44w container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:59 crc kubenswrapper[4824]: I1209 11:13:59.689397 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" podUID="93c989b8-7b86-4339-bbab-5886c7d13dc9" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.55:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:13:59 crc kubenswrapper[4824]: I1209 11:13:59.851890 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-bvp6v container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:13:59 crc kubenswrapper[4824]: I1209 11:13:59.852250 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" podUID="858d899b-800f-4639-8fbd-4f1ccad44991" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:00 crc kubenswrapper[4824]: I1209 11:14:00.196421 4824 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Liveness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:00 crc kubenswrapper[4824]: I1209 11:14:00.196499 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:00 crc kubenswrapper[4824]: I1209 11:14:00.387207 4824 patch_prober.go:28] interesting pod/thanos-querier-7fc6d7f97-nsq7f container/kube-rbac-proxy-web namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.72:9091/-/healthy\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:00 crc kubenswrapper[4824]: I1209 11:14:00.387267 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" podUID="d030e6a9-d47a-47bf-9c24-ff0ca58f71e6" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.72:9091/-/healthy\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:00 crc kubenswrapper[4824]: I1209 11:14:00.387398 4824 patch_prober.go:28] interesting pod/thanos-querier-7fc6d7f97-nsq7f container/kube-rbac-proxy-web namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.72:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:00 crc kubenswrapper[4824]: I1209 11:14:00.387454 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" podUID="d030e6a9-d47a-47bf-9c24-ff0ca58f71e6" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.72:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:00 crc kubenswrapper[4824]: I1209 11:14:00.509987 4824 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-qn2qz container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.5:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:00 crc kubenswrapper[4824]: I1209 11:14:00.510048 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" podUID="612ebf47-a6c4-4a62-8d97-91a003f49c44" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.5:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:00 crc kubenswrapper[4824]: I1209 11:14:00.510082 4824 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-qn2qz container/operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.5:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:00 crc kubenswrapper[4824]: I1209 11:14:00.510116 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" podUID="612ebf47-a6c4-4a62-8d97-91a003f49c44" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.5:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:00 crc kubenswrapper[4824]: I1209 11:14:00.510183 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-controller-manager-69bb78d57c-p4l9h" podUID="2fed244b-a135-4d95-a5e5-2eea4bcde7cb" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.94:8080/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:00 crc kubenswrapper[4824]: I1209 11:14:00.746663 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="73d6bd70-44c7-4eed-a93a-36df636869cf" containerName="galera" probeResult="failure" output="command timed out" Dec 09 11:14:00 crc kubenswrapper[4824]: I1209 11:14:00.746668 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="73d6bd70-44c7-4eed-a93a-36df636869cf" containerName="galera" probeResult="failure" output="command timed out" Dec 09 11:14:00 crc kubenswrapper[4824]: I1209 11:14:00.762999 4824 patch_prober.go:28] interesting pod/perses-operator-5446b9c989-pkmss container/perses-operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.30:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:00 crc kubenswrapper[4824]: I1209 11:14:00.763072 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/perses-operator-5446b9c989-pkmss" podUID="2ecf9e90-37fe-416d-9e7a-400373cfbc8d" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.30:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:00 crc kubenswrapper[4824]: I1209 11:14:00.804092 4824 patch_prober.go:28] interesting pod/perses-operator-5446b9c989-pkmss container/perses-operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.30:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:00 crc kubenswrapper[4824]: I1209 11:14:00.804151 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/perses-operator-5446b9c989-pkmss" podUID="2ecf9e90-37fe-416d-9e7a-400373cfbc8d" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.30:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:00 crc kubenswrapper[4824]: I1209 11:14:00.845016 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" podUID="bab491e7-4eec-4ba1-975d-fb2468372ade" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.95:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:00 crc kubenswrapper[4824]: I1209 11:14:00.887049 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" podUID="bab491e7-4eec-4ba1-975d-fb2468372ade" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.95:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:01 crc kubenswrapper[4824]: I1209 11:14:01.132622 4824 patch_prober.go:28] interesting pod/controller-manager-69949c7bd4-pmmk2 container/controller-manager namespace/openshift-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.78:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:01 crc kubenswrapper[4824]: I1209 11:14:01.132706 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" podUID="ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.78:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:01 crc kubenswrapper[4824]: I1209 11:14:01.132810 4824 patch_prober.go:28] interesting pod/controller-manager-69949c7bd4-pmmk2 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.78:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:01 crc kubenswrapper[4824]: I1209 11:14:01.132836 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" podUID="ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.78:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:01 crc kubenswrapper[4824]: I1209 11:14:01.668983 4824 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4tt6w container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.56:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:01 crc kubenswrapper[4824]: I1209 11:14:01.668987 4824 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4tt6w container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.56:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:01 crc kubenswrapper[4824]: I1209 11:14:01.669050 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4tt6w" podUID="1b70aae4-4b6b-457c-a358-0f7ce11b8206" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.56:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:01 crc kubenswrapper[4824]: I1209 11:14:01.669244 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-4tt6w" podUID="1b70aae4-4b6b-457c-a358-0f7ce11b8206" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.56:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:01 crc kubenswrapper[4824]: I1209 11:14:01.744584 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-northd-0" podUID="c2dd0680-d32f-45e8-b834-6a327d2eaa21" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 09 11:14:01 crc kubenswrapper[4824]: I1209 11:14:01.745395 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="c2dd0680-d32f-45e8-b834-6a327d2eaa21" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 09 11:14:01 crc kubenswrapper[4824]: I1209 11:14:01.746929 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="844884cd-6813-4fa2-88e4-7a5994bbd3cb" containerName="prometheus" probeResult="failure" output="command timed out" Dec 09 11:14:01 crc kubenswrapper[4824]: I1209 11:14:01.747465 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-k8s-0" podUID="844884cd-6813-4fa2-88e4-7a5994bbd3cb" containerName="prometheus" probeResult="failure" output="command timed out" Dec 09 11:14:01 crc kubenswrapper[4824]: I1209 11:14:01.794030 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-ntx24" podUID="ad8f524a-8c95-47ab-b74d-9f83331fce76" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:01 crc kubenswrapper[4824]: I1209 11:14:01.794026 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-ntx24" podUID="ad8f524a-8c95-47ab-b74d-9f83331fce76" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:01 crc kubenswrapper[4824]: I1209 11:14:01.794181 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-ntx24" podUID="ad8f524a-8c95-47ab-b74d-9f83331fce76" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:01 crc kubenswrapper[4824]: I1209 11:14:01.835230 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6mfg8" podUID="5b34da01-28d8-476c-a2a5-a098489d90c1" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.102:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:01 crc kubenswrapper[4824]: I1209 11:14:01.919184 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/controller-f8648f98b-wxgt4" podUID="3af18cd9-4124-4f24-83ea-0ad26ad9ae43" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.97:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:01 crc kubenswrapper[4824]: I1209 11:14:01.919404 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/controller-f8648f98b-wxgt4" podUID="3af18cd9-4124-4f24-83ea-0ad26ad9ae43" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.97:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:01 crc kubenswrapper[4824]: I1209 11:14:01.960207 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4" podUID="143e48cd-c956-448c-8dcb-5858a582e29c" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.104:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:02 crc kubenswrapper[4824]: I1209 11:14:02.074009 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-bf7pj" podUID="ccc192e8-2648-4b1b-8420-1cbd1a27a916" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:02 crc kubenswrapper[4824]: I1209 11:14:02.074016 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-k92kx" podUID="30f20f09-f33f-4c7a-a4b4-41b51f73d692" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.107:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:02 crc kubenswrapper[4824]: I1209 11:14:02.110931 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/prometheus-metric-storage-0" podUID="fb8f1e44-4c21-423a-bacc-1cca7d7715c3" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.167:9090/-/healthy\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:02 crc kubenswrapper[4824]: I1209 11:14:02.115025 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9z2kx" podUID="d1323625-eb0f-4606-bee7-7acbd453d4a5" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.111:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:02 crc kubenswrapper[4824]: I1209 11:14:02.115846 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="fb8f1e44-4c21-423a-bacc-1cca7d7715c3" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.167:9090/-/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:02 crc kubenswrapper[4824]: I1209 11:14:02.247022 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-967d97867-lfprd" podUID="b0c20e13-b233-4b21-8f47-facef9603735" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:02 crc kubenswrapper[4824]: I1209 11:14:02.294114 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4q2h7" podUID="7de1c7e2-adcb-4311-926c-ffc42e3e5fde" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.110:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:02 crc kubenswrapper[4824]: I1209 11:14:02.335071 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-8rvmd" podUID="bd655e5b-2ac3-4fe8-ad58-6012d0180897" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.96:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:02 crc kubenswrapper[4824]: I1209 11:14:02.376078 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-8rvmd" podUID="bd655e5b-2ac3-4fe8-ad58-6012d0180897" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.96:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:02 crc kubenswrapper[4824]: I1209 11:14:02.416094 4824 patch_prober.go:28] interesting pod/console-operator-58897d9998-wgtx2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:02 crc kubenswrapper[4824]: I1209 11:14:02.416175 4824 patch_prober.go:28] interesting pod/console-operator-58897d9998-wgtx2 container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:02 crc kubenswrapper[4824]: I1209 11:14:02.416264 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" podUID="cd6c14a0-99d8-485d-81f3-cc4ebe5a943b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:02 crc kubenswrapper[4824]: I1209 11:14:02.416181 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" podUID="cd6c14a0-99d8-485d-81f3-cc4ebe5a943b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:02 crc kubenswrapper[4824]: I1209 11:14:02.528019 4824 patch_prober.go:28] interesting pod/metrics-server-65846dd9c4-g8tlp container/metrics-server namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.79:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:02 crc kubenswrapper[4824]: I1209 11:14:02.528039 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl" podUID="b6577050-6a21-43bb-84aa-20aad247aafc" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:02 crc kubenswrapper[4824]: I1209 11:14:02.528105 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" podUID="e153e30f-4f5e-4a38-8efb-6452096e25d7" containerName="metrics-server" probeResult="failure" output="Get \"https://10.217.0.79:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:02 crc kubenswrapper[4824]: I1209 11:14:02.528160 4824 patch_prober.go:28] interesting pod/metrics-server-65846dd9c4-g8tlp container/metrics-server namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.79:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:02 crc kubenswrapper[4824]: I1209 11:14:02.528246 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" podUID="e153e30f-4f5e-4a38-8efb-6452096e25d7" containerName="metrics-server" probeResult="failure" output="Get \"https://10.217.0.79:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:02 crc kubenswrapper[4824]: I1209 11:14:02.808003 4824 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-bxplc container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:02 crc kubenswrapper[4824]: I1209 11:14:02.808071 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" podUID="c3400b7c-5666-4b36-a50f-7800f6737527" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:02 crc kubenswrapper[4824]: I1209 11:14:02.873029 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8" podUID="23621a98-1d6b-421a-976c-965cecfe6db1" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.113:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:02 crc kubenswrapper[4824]: I1209 11:14:02.913993 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" podUID="d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:03 crc kubenswrapper[4824]: I1209 11:14:03.001087 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" podUID="383a67da-0f83-41e2-82d6-809d1b6cdbec" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.118:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:03 crc kubenswrapper[4824]: I1209 11:14:03.001185 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh" podUID="91bf6cd9-a4fe-46e8-b1e7-39e57c1398b3" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:03 crc kubenswrapper[4824]: I1209 11:14:03.084388 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9" podUID="31749799-9eb4-403f-a61d-9d50d8bc8367" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.117:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:03 crc kubenswrapper[4824]: I1209 11:14:03.085577 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl" podUID="6e91c44f-6221-4dfd-9c18-2bbedff02850" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.119:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:03 crc kubenswrapper[4824]: I1209 11:14:03.126867 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" podUID="feab57e6-788b-4f23-9e99-aa248786052d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:03 crc kubenswrapper[4824]: I1209 11:14:03.210223 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" podUID="bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.121:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:03 crc kubenswrapper[4824]: I1209 11:14:03.210714 4824 patch_prober.go:28] interesting pod/monitoring-plugin-6f74c7bb76-pbw94 container/monitoring-plugin namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.80:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:03 crc kubenswrapper[4824]: I1209 11:14:03.210932 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/monitoring-plugin-6f74c7bb76-pbw94" podUID="364b1858-6a46-4db4-9d9d-0ec19d54abc9" containerName="monitoring-plugin" probeResult="failure" output="Get \"https://10.217.0.80:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:03 crc kubenswrapper[4824]: I1209 11:14:03.211276 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-q7psb" podUID="0af0249c-ed71-4f92-b1a3-cb6f43b92529" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.122:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:03 crc kubenswrapper[4824]: I1209 11:14:03.316045 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" podUID="9f87669d-33a9-4269-808c-ca6c718f762c" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.44:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:03 crc kubenswrapper[4824]: I1209 11:14:03.316058 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" podUID="9f87669d-33a9-4269-808c-ca6c718f762c" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.44:6080/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:03 crc kubenswrapper[4824]: I1209 11:14:03.401025 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-5xfv2" podUID="87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:03 crc kubenswrapper[4824]: I1209 11:14:03.401028 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-5xfv2" podUID="87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:03 crc kubenswrapper[4824]: I1209 11:14:03.551900 4824 patch_prober.go:28] interesting pod/logging-loki-distributor-76cc67bf56-nl2vt container/loki-distributor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.51:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:03 crc kubenswrapper[4824]: I1209 11:14:03.551967 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" podUID="72ff7b94-63fd-45d1-a803-bc1ccf4388ed" containerName="loki-distributor" probeResult="failure" output="Get \"https://10.217.0.51:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.277915 4824 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-9c2mr container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.278245 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" podUID="62823972-71dc-4f7d-b4f6-da0cb90dcf36" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.294163 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.294293 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.328386 4824 patch_prober.go:28] interesting pod/router-default-5444994796-mj7jx container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.328482 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.494095 4824 patch_prober.go:28] interesting pod/logging-loki-querier-5895d59bb8-89dgf container/loki-querier namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.52:3101/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.494181 4824 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-9c2mr container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.494095 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-b9bnh container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.494226 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-b9bnh" podUID="6efb8245-07e9-4c31-b8ab-53ab3685e593" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.494236 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" podUID="62823972-71dc-4f7d-b4f6-da0cb90dcf36" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.494178 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" podUID="1006b7c9-7244-44af-8bc9-52787d891f7f" containerName="loki-querier" probeResult="failure" output="Get \"https://10.217.0.52:3101/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.494613 4824 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-c2dfm container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.494631 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c2dfm" podUID="a7c02c2b-70ae-4b6f-85ae-5a38b737a73b" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.494655 4824 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-c2dfm container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.494666 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c2dfm" podUID="a7c02c2b-70ae-4b6f-85ae-5a38b737a73b" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.494801 4824 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-tg5lv container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.494873 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" podUID="45732d49-ce14-439b-8b7a-fb9fdf267fb1" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.494932 4824 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-tg5lv container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.494968 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" podUID="45732d49-ce14-439b-8b7a-fb9fdf267fb1" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.535997 4824 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-k8bgl container/package-server-manager namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.536061 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" podUID="6c10edf9-cf75-4633-8d4c-9bbb9d9b6339" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.536246 4824 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-k8bgl container/package-server-manager namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.536262 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" podUID="6c10edf9-cf75-4633-8d4c-9bbb9d9b6339" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.536319 4824 patch_prober.go:28] interesting pod/router-default-5444994796-mj7jx container/router namespace/openshift-ingress: Liveness probe status=failure output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.536378 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.536441 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-b9bnh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.536467 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b9bnh" podUID="6efb8245-07e9-4c31-b8ab-53ab3685e593" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.560413 4824 patch_prober.go:28] interesting pod/logging-loki-query-frontend-84558f7c9f-9d5fm container/loki-query-frontend namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:04 crc kubenswrapper[4824]: I1209 11:14:04.560505 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" podUID="e05c11bd-cd92-4bac-adea-0a6049ccfb39" containerName="loki-query-frontend" probeResult="failure" output="Get \"https://10.217.0.53:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.276647 4824 patch_prober.go:28] interesting pod/oauth-openshift-f578d5c8f-5jzzv container/oauth-openshift namespace/openshift-authentication: Liveness probe status=failure output="Get \"https://10.217.0.62:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.276717 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" podUID="3d1e2cd1-5fc5-4a76-b1ff-07db11b07708" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.62:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.281062 4824 patch_prober.go:28] interesting pod/console-7c85c585b4-tgrff container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.139:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.281313 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-7c85c585b4-tgrff" podUID="e61278e7-d0a8-4039-909d-c2812c8a4a81" containerName="console" probeResult="failure" output="Get \"https://10.217.0.139:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.282462 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-vz44w container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:8081/ready\": context deadline exceeded" start-of-body= Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.282528 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" podUID="93c989b8-7b86-4339-bbab-5886c7d13dc9" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.55:8081/ready\": context deadline exceeded" Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.282821 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-vz44w container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:8083/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.282926 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" podUID="93c989b8-7b86-4339-bbab-5886c7d13dc9" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.55:8083/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.293955 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-bvp6v container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.293994 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" podUID="858d899b-800f-4639-8fbd-4f1ccad44991" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.302209 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-bvp6v container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.302275 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" podUID="858d899b-800f-4639-8fbd-4f1ccad44991" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.54:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.304412 4824 patch_prober.go:28] interesting pod/oauth-openshift-f578d5c8f-5jzzv container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.62:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.304443 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" podUID="3d1e2cd1-5fc5-4a76-b1ff-07db11b07708" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.62:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.319650 4824 patch_prober.go:28] interesting pod/logging-loki-compactor-0 container/loki-compactor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.65:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.319738 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-compactor-0" podUID="379b9136-1cdf-4786-8494-e99c7161b202" containerName="loki-compactor" probeResult="failure" output="Get \"https://10.217.0.65:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.374506 4824 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.64:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.374584 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="9d24f296-729f-4eec-a1ae-d6b904399394" containerName="loki-ingester" probeResult="failure" output="Get \"https://10.217.0.64:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.386939 4824 patch_prober.go:28] interesting pod/thanos-querier-7fc6d7f97-nsq7f container/kube-rbac-proxy-web namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.72:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.387013 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" podUID="d030e6a9-d47a-47bf-9c24-ff0ca58f71e6" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.72:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.455192 4824 patch_prober.go:28] interesting pod/logging-loki-index-gateway-0 container/loki-index-gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.74:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.455264 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-index-gateway-0" podUID="6baf0bfe-7be6-4309-93da-e86174e4654e" containerName="loki-index-gateway" probeResult="failure" output="Get \"https://10.217.0.74:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.761844 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-index-2jnrg" podUID="25c6995d-6a34-43fc-9b1b-6c34ed9aec46" containerName="registry-server" probeResult="failure" output="command timed out" Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.761840 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-index-2jnrg" podUID="25c6995d-6a34-43fc-9b1b-6c34ed9aec46" containerName="registry-server" probeResult="failure" output="command timed out" Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.950077 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" podUID="e7e496dc-23c4-47cd-9c5a-aa4430130849" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.123:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:05 crc kubenswrapper[4824]: I1209 11:14:05.950204 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" podUID="e7e496dc-23c4-47cd-9c5a-aa4430130849" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.123:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:06 crc kubenswrapper[4824]: I1209 11:14:06.074062 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" podUID="c0ac631f-5d13-4814-84d3-61f7bd704f9e" containerName="hostpath-provisioner" probeResult="failure" output="Get \"http://10.217.0.43:9898/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:06 crc kubenswrapper[4824]: I1209 11:14:06.745773 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-northd-0" podUID="c2dd0680-d32f-45e8-b834-6a327d2eaa21" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 09 11:14:06 crc kubenswrapper[4824]: I1209 11:14:06.745761 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="c2dd0680-d32f-45e8-b834-6a327d2eaa21" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 09 11:14:06 crc kubenswrapper[4824]: I1209 11:14:06.763530 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ovn-northd-0" Dec 09 11:14:06 crc kubenswrapper[4824]: I1209 11:14:06.766185 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 09 11:14:06 crc kubenswrapper[4824]: I1209 11:14:06.785103 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ovn-northd" containerStatusID={"Type":"cri-o","ID":"a13a50a9012203c512c34fec894886dfe2d054768904caa6f9ab13fa1fc56c87"} pod="openstack/ovn-northd-0" containerMessage="Container ovn-northd failed liveness probe, will be restarted" Dec 09 11:14:06 crc kubenswrapper[4824]: I1209 11:14:06.786284 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="c2dd0680-d32f-45e8-b834-6a327d2eaa21" containerName="ovn-northd" containerID="cri-o://a13a50a9012203c512c34fec894886dfe2d054768904caa6f9ab13fa1fc56c87" gracePeriod=30 Dec 09 11:14:07 crc kubenswrapper[4824]: I1209 11:14:07.054738 4824 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:07 crc kubenswrapper[4824]: I1209 11:14:07.054834 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:07 crc kubenswrapper[4824]: I1209 11:14:07.066624 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/kube-state-metrics-0" podUID="54621281-db00-41e4-b617-032435893391" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.0.254:8080/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:07 crc kubenswrapper[4824]: I1209 11:14:07.066959 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="54621281-db00-41e4-b617-032435893391" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.0.254:8081/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:07 crc kubenswrapper[4824]: I1209 11:14:07.251238 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-vsdpm" podUID="c0ac631f-5d13-4814-84d3-61f7bd704f9e" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:14:07 crc kubenswrapper[4824]: I1209 11:14:07.680553 4824 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-j4zxt container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:07 crc kubenswrapper[4824]: I1209 11:14:07.680642 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" podUID="7b26f121-04f6-4501-84b2-1833b927aa14" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:07 crc kubenswrapper[4824]: I1209 11:14:07.680722 4824 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-j4zxt container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:07 crc kubenswrapper[4824]: I1209 11:14:07.680740 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" podUID="7b26f121-04f6-4501-84b2-1833b927aa14" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:07 crc kubenswrapper[4824]: I1209 11:14:07.743536 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-k8s-0" podUID="844884cd-6813-4fa2-88e4-7a5994bbd3cb" containerName="prometheus" probeResult="failure" output="command timed out" Dec 09 11:14:07 crc kubenswrapper[4824]: I1209 11:14:07.745406 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="844884cd-6813-4fa2-88e4-7a5994bbd3cb" containerName="prometheus" probeResult="failure" output="command timed out" Dec 09 11:14:07 crc kubenswrapper[4824]: I1209 11:14:07.746224 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="c2dd0680-d32f-45e8-b834-6a327d2eaa21" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 09 11:14:08 crc kubenswrapper[4824]: I1209 11:14:08.021289 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" podUID="1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.108:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:08 crc kubenswrapper[4824]: I1209 11:14:08.063176 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" podUID="1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.108:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:08 crc kubenswrapper[4824]: I1209 11:14:08.299887 4824 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-4dlng container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:08 crc kubenswrapper[4824]: I1209 11:14:08.300277 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" podUID="86c42164-acb8-4a27-9a0f-48307cd304e4" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:08 crc kubenswrapper[4824]: I1209 11:14:08.299906 4824 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-4dlng container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:08 crc kubenswrapper[4824]: I1209 11:14:08.300488 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" podUID="86c42164-acb8-4a27-9a0f-48307cd304e4" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:08 crc kubenswrapper[4824]: I1209 11:14:08.638126 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" podUID="c2889de3-49b2-4465-8235-37f9e58b42a3" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.116:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:08 crc kubenswrapper[4824]: I1209 11:14:08.638115 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" podUID="c2889de3-49b2-4465-8235-37f9e58b42a3" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.116:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:08 crc kubenswrapper[4824]: I1209 11:14:08.749349 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-operators-582xj" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" containerName="registry-server" probeResult="failure" output="command timed out" Dec 09 11:14:08 crc kubenswrapper[4824]: I1209 11:14:08.749468 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-operators-582xj" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" containerName="registry-server" probeResult="failure" output="command timed out" Dec 09 11:14:08 crc kubenswrapper[4824]: I1209 11:14:08.749521 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-marketplace-95jfr" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" containerName="registry-server" probeResult="failure" output="command timed out" Dec 09 11:14:08 crc kubenswrapper[4824]: I1209 11:14:08.749601 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-95jfr" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" containerName="registry-server" probeResult="failure" output="command timed out" Dec 09 11:14:09 crc kubenswrapper[4824]: I1209 11:14:09.305394 4824 patch_prober.go:28] interesting pod/route-controller-manager-7f84c5b8f9-rqmr9 container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:09 crc kubenswrapper[4824]: I1209 11:14:09.306138 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" podUID="afe92d2b-bce3-445e-9c26-9c533353e7a3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:09 crc kubenswrapper[4824]: I1209 11:14:09.305808 4824 patch_prober.go:28] interesting pod/route-controller-manager-7f84c5b8f9-rqmr9 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:09 crc kubenswrapper[4824]: I1209 11:14:09.306428 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" podUID="afe92d2b-bce3-445e-9c26-9c533353e7a3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:09 crc kubenswrapper[4824]: I1209 11:14:09.461656 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-operator-b846b8d48-dhtnn" podUID="ebb1c8e1-3960-4c38-969b-c483143c87aa" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.101:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:09 crc kubenswrapper[4824]: I1209 11:14:09.463276 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-controller-operator-b846b8d48-dhtnn" podUID="ebb1c8e1-3960-4c38-969b-c483143c87aa" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.101:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:09 crc kubenswrapper[4824]: I1209 11:14:09.688541 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-vz44w container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:09 crc kubenswrapper[4824]: I1209 11:14:09.688622 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" podUID="93c989b8-7b86-4339-bbab-5886c7d13dc9" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.55:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:09 crc kubenswrapper[4824]: I1209 11:14:09.754046 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="56148e0f-636f-410e-bfb5-342da01e8c76" containerName="galera" probeResult="failure" output="command timed out" Dec 09 11:14:09 crc kubenswrapper[4824]: I1209 11:14:09.754046 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="56148e0f-636f-410e-bfb5-342da01e8c76" containerName="galera" probeResult="failure" output="command timed out" Dec 09 11:14:09 crc kubenswrapper[4824]: I1209 11:14:09.765399 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 09 11:14:09 crc kubenswrapper[4824]: I1209 11:14:09.766846 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/openstack-galera-0" Dec 09 11:14:09 crc kubenswrapper[4824]: I1209 11:14:09.768519 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 09 11:14:09 crc kubenswrapper[4824]: I1209 11:14:09.768546 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Dec 09 11:14:09 crc kubenswrapper[4824]: I1209 11:14:09.774011 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="galera" containerStatusID={"Type":"cri-o","ID":"4f85804d9b180ab6d108fd3769e0a57f89c384076fef52095c519bf02c5c868e"} pod="openstack/openstack-galera-0" containerMessage="Container galera failed liveness probe, will be restarted" Dec 09 11:14:09 crc kubenswrapper[4824]: I1209 11:14:09.774006 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-central-agent" containerStatusID={"Type":"cri-o","ID":"c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-central-agent failed liveness probe, will be restarted" Dec 09 11:14:09 crc kubenswrapper[4824]: I1209 11:14:09.774895 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" containerName="ceilometer-central-agent" containerID="cri-o://c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" gracePeriod=30 Dec 09 11:14:09 crc kubenswrapper[4824]: I1209 11:14:09.851249 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-bvp6v container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:09 crc kubenswrapper[4824]: I1209 11:14:09.851348 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" podUID="858d899b-800f-4639-8fbd-4f1ccad44991" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:10 crc kubenswrapper[4824]: I1209 11:14:10.062382 4824 trace.go:236] Trace[1032927238]: "Calculate volume metrics of storage for pod openshift-logging/logging-loki-ingester-0" (09-Dec-2025 11:14:02.965) (total time: 7095ms): Dec 09 11:14:10 crc kubenswrapper[4824]: Trace[1032927238]: [7.095202485s] [7.095202485s] END Dec 09 11:14:10 crc kubenswrapper[4824]: I1209 11:14:10.062382 4824 trace.go:236] Trace[280648837]: "Calculate volume metrics of persistence for pod openstack/rabbitmq-server-1" (09-Dec-2025 11:14:04.280) (total time: 5780ms): Dec 09 11:14:10 crc kubenswrapper[4824]: Trace[280648837]: [5.780118793s] [5.780118793s] END Dec 09 11:14:10 crc kubenswrapper[4824]: E1209 11:14:10.107579 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a13a50a9012203c512c34fec894886dfe2d054768904caa6f9ab13fa1fc56c87" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 09 11:14:10 crc kubenswrapper[4824]: E1209 11:14:10.111078 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a13a50a9012203c512c34fec894886dfe2d054768904caa6f9ab13fa1fc56c87" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 09 11:14:10 crc kubenswrapper[4824]: E1209 11:14:10.112742 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a13a50a9012203c512c34fec894886dfe2d054768904caa6f9ab13fa1fc56c87" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 09 11:14:10 crc kubenswrapper[4824]: E1209 11:14:10.112840 4824 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="c2dd0680-d32f-45e8-b834-6a327d2eaa21" containerName="ovn-northd" Dec 09 11:14:10 crc kubenswrapper[4824]: I1209 11:14:10.195399 4824 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Liveness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:10 crc kubenswrapper[4824]: I1209 11:14:10.195481 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:10 crc kubenswrapper[4824]: I1209 11:14:10.280993 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/certified-operators-2cbmd" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" containerName="registry-server" probeResult="failure" output=< Dec 09 11:14:10 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 11:14:10 crc kubenswrapper[4824]: > Dec 09 11:14:10 crc kubenswrapper[4824]: I1209 11:14:10.281046 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/certified-operators-2cbmd" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" containerName="registry-server" probeResult="failure" output=< Dec 09 11:14:10 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 11:14:10 crc kubenswrapper[4824]: > Dec 09 11:14:10 crc kubenswrapper[4824]: I1209 11:14:10.280993 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-jm5ld" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" containerName="registry-server" probeResult="failure" output=< Dec 09 11:14:10 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 11:14:10 crc kubenswrapper[4824]: > Dec 09 11:14:10 crc kubenswrapper[4824]: I1209 11:14:10.282207 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/community-operators-jm5ld" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" containerName="registry-server" probeResult="failure" output=< Dec 09 11:14:10 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 11:14:10 crc kubenswrapper[4824]: > Dec 09 11:14:10 crc kubenswrapper[4824]: I1209 11:14:10.428505 4824 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-qn2qz container/operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.5:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:10 crc kubenswrapper[4824]: I1209 11:14:10.428624 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" podUID="612ebf47-a6c4-4a62-8d97-91a003f49c44" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.5:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:10 crc kubenswrapper[4824]: I1209 11:14:10.469764 4824 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-qn2qz container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.5:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:10 crc kubenswrapper[4824]: I1209 11:14:10.469898 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" podUID="612ebf47-a6c4-4a62-8d97-91a003f49c44" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.5:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:10 crc kubenswrapper[4824]: I1209 11:14:10.511165 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-controller-manager-69bb78d57c-p4l9h" podUID="2fed244b-a135-4d95-a5e5-2eea4bcde7cb" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.94:8080/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:10 crc kubenswrapper[4824]: I1209 11:14:10.680264 4824 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-j4zxt container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:10 crc kubenswrapper[4824]: I1209 11:14:10.680338 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" podUID="7b26f121-04f6-4501-84b2-1833b927aa14" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:10 crc kubenswrapper[4824]: I1209 11:14:10.680344 4824 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-j4zxt container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:10 crc kubenswrapper[4824]: I1209 11:14:10.680412 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" podUID="7b26f121-04f6-4501-84b2-1833b927aa14" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:10 crc kubenswrapper[4824]: I1209 11:14:10.744485 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="73d6bd70-44c7-4eed-a93a-36df636869cf" containerName="galera" probeResult="failure" output="command timed out" Dec 09 11:14:10 crc kubenswrapper[4824]: I1209 11:14:10.744600 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="56148e0f-636f-410e-bfb5-342da01e8c76" containerName="galera" probeResult="failure" output="command timed out" Dec 09 11:14:10 crc kubenswrapper[4824]: I1209 11:14:10.747176 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="73d6bd70-44c7-4eed-a93a-36df636869cf" containerName="galera" probeResult="failure" output="command timed out" Dec 09 11:14:10 crc kubenswrapper[4824]: I1209 11:14:10.765607 4824 patch_prober.go:28] interesting pod/perses-operator-5446b9c989-pkmss container/perses-operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.30:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:10 crc kubenswrapper[4824]: I1209 11:14:10.765740 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/perses-operator-5446b9c989-pkmss" podUID="2ecf9e90-37fe-416d-9e7a-400373cfbc8d" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.30:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:10 crc kubenswrapper[4824]: I1209 11:14:10.900286 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" podUID="bab491e7-4eec-4ba1-975d-fb2468372ade" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.95:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:10 crc kubenswrapper[4824]: I1209 11:14:10.901110 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" podUID="bab491e7-4eec-4ba1-975d-fb2468372ade" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.95:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:11 crc kubenswrapper[4824]: I1209 11:14:11.133680 4824 patch_prober.go:28] interesting pod/controller-manager-69949c7bd4-pmmk2 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.78:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:11 crc kubenswrapper[4824]: I1209 11:14:11.133682 4824 patch_prober.go:28] interesting pod/controller-manager-69949c7bd4-pmmk2 container/controller-manager namespace/openshift-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.78:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:11 crc kubenswrapper[4824]: I1209 11:14:11.133746 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" podUID="ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.78:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:11 crc kubenswrapper[4824]: I1209 11:14:11.133856 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" podUID="ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.78:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:11 crc kubenswrapper[4824]: I1209 11:14:11.668978 4824 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4tt6w container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.56:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:11 crc kubenswrapper[4824]: I1209 11:14:11.669351 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-4tt6w" podUID="1b70aae4-4b6b-457c-a358-0f7ce11b8206" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.56:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:11 crc kubenswrapper[4824]: I1209 11:14:11.669087 4824 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4tt6w container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.56:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:11 crc kubenswrapper[4824]: I1209 11:14:11.669421 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4tt6w" podUID="1b70aae4-4b6b-457c-a358-0f7ce11b8206" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.56:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:11 crc kubenswrapper[4824]: I1209 11:14:11.709957 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-ntx24" podUID="ad8f524a-8c95-47ab-b74d-9f83331fce76" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:11 crc kubenswrapper[4824]: I1209 11:14:11.791958 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-ntx24" podUID="ad8f524a-8c95-47ab-b74d-9f83331fce76" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:11 crc kubenswrapper[4824]: I1209 11:14:11.791966 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-ntx24" podUID="ad8f524a-8c95-47ab-b74d-9f83331fce76" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:11 crc kubenswrapper[4824]: I1209 11:14:11.878138 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6mfg8" podUID="5b34da01-28d8-476c-a2a5-a098489d90c1" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.102:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:11 crc kubenswrapper[4824]: I1209 11:14:11.961913 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-lpf4s" podUID="270a1cf5-f943-4749-8c9a-1c6adf3c45b4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.103:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.043986 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6mfg8" podUID="5b34da01-28d8-476c-a2a5-a098489d90c1" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.102:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.043995 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/controller-f8648f98b-wxgt4" podUID="3af18cd9-4124-4f24-83ea-0ad26ad9ae43" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.97:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.058745 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6mfg8" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.086385 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4" podUID="143e48cd-c956-448c-8dcb-5858a582e29c" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.104:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.086608 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.168031 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vx285" podUID="54693584-08f2-4326-8062-92d68899d7e4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.105:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.168635 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-lpf4s" podUID="270a1cf5-f943-4749-8c9a-1c6adf3c45b4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.103:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.335940 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/controller-f8648f98b-wxgt4" podUID="3af18cd9-4124-4f24-83ea-0ad26ad9ae43" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.97:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.335970 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-bf7pj" podUID="ccc192e8-2648-4b1b-8420-1cbd1a27a916" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.418178 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9z2kx" podUID="d1323625-eb0f-4606-bee7-7acbd453d4a5" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.111:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.418550 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4" podUID="143e48cd-c956-448c-8dcb-5858a582e29c" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.104:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.419069 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/prometheus-metric-storage-0" podUID="fb8f1e44-4c21-423a-bacc-1cca7d7715c3" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.167:9090/-/healthy\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.419146 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="fb8f1e44-4c21-423a-bacc-1cca7d7715c3" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.167:9090/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.419175 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vx285" podUID="54693584-08f2-4326-8062-92d68899d7e4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.105:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.502079 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/ironic-operator-controller-manager-967d97867-lfprd" podUID="b0c20e13-b233-4b21-8f47-facef9603735" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.502177 4824 patch_prober.go:28] interesting pod/nmstate-webhook-5f6d4c5ccb-jklpr container/nmstate-webhook namespace/openshift-nmstate: Readiness probe status=failure output="Get \"https://10.217.0.88:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.502243 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jklpr" podUID="0b932274-5eba-4e0d-8a64-1f469a6ab3d1" containerName="nmstate-webhook" probeResult="failure" output="Get \"https://10.217.0.88:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.502266 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-bf7pj" podUID="ccc192e8-2648-4b1b-8420-1cbd1a27a916" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.502571 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-k92kx" podUID="30f20f09-f33f-4c7a-a4b4-41b51f73d692" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.107:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.626080 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-8rvmd" podUID="bd655e5b-2ac3-4fe8-ad58-6012d0180897" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.96:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.666971 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-8rvmd" podUID="bd655e5b-2ac3-4fe8-ad58-6012d0180897" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.96:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.667029 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-k92kx" podUID="30f20f09-f33f-4c7a-a4b4-41b51f73d692" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.107:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.667092 4824 patch_prober.go:28] interesting pod/console-operator-58897d9998-wgtx2 container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.667097 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-9z2kx" podUID="d1323625-eb0f-4606-bee7-7acbd453d4a5" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.111:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.667142 4824 patch_prober.go:28] interesting pod/console-operator-58897d9998-wgtx2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.667161 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" podUID="cd6c14a0-99d8-485d-81f3-cc4ebe5a943b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.667118 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" podUID="cd6c14a0-99d8-485d-81f3-cc4ebe5a943b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.667236 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.667265 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.667655 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-967d97867-lfprd" podUID="b0c20e13-b233-4b21-8f47-facef9603735" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.667806 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-967d97867-lfprd" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.669769 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="console-operator" containerStatusID={"Type":"cri-o","ID":"97c7e3b1f76efb699d4015e9514f819e32272fc0f7184743e1eb1a7cea87380e"} pod="openshift-console-operator/console-operator-58897d9998-wgtx2" containerMessage="Container console-operator failed liveness probe, will be restarted" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.670227 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" podUID="cd6c14a0-99d8-485d-81f3-cc4ebe5a943b" containerName="console-operator" containerID="cri-o://97c7e3b1f76efb699d4015e9514f819e32272fc0f7184743e1eb1a7cea87380e" gracePeriod=30 Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.708109 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl" podUID="b6577050-6a21-43bb-84aa-20aad247aafc" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:12 crc kubenswrapper[4824]: I1209 11:14:12.708248 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:12.744378 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-handler-5strj" podUID="952171d0-1a2e-4801-bc11-4012ba19588c" containerName="nmstate-handler" probeResult="failure" output="command timed out" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:12.746107 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-k8s-0" podUID="844884cd-6813-4fa2-88e4-7a5994bbd3cb" containerName="prometheus" probeResult="failure" output="command timed out" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:12.746107 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="844884cd-6813-4fa2-88e4-7a5994bbd3cb" containerName="prometheus" probeResult="failure" output="command timed out" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:12.746282 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-0" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:12.749008 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl" podUID="b6577050-6a21-43bb-84aa-20aad247aafc" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:12.749065 4824 patch_prober.go:28] interesting pod/metrics-server-65846dd9c4-g8tlp container/metrics-server namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.79:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:12.749094 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" podUID="e153e30f-4f5e-4a38-8efb-6452096e25d7" containerName="metrics-server" probeResult="failure" output="Get \"https://10.217.0.79:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:12.749129 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:12.749147 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4q2h7" podUID="7de1c7e2-adcb-4311-926c-ffc42e3e5fde" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.110:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:12.749229 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4q2h7" podUID="7de1c7e2-adcb-4311-926c-ffc42e3e5fde" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.110:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:12.766748 4824 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-bxplc container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:12.766844 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" podUID="c3400b7c-5666-4b36-a50f-7800f6737527" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:12.766898 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:12.767917 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="authentication-operator" containerStatusID={"Type":"cri-o","ID":"ededff8a2f4ee4e72db8028a40598edd0ce0981b53e5b3d8e13b10053bf64cd1"} pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" containerMessage="Container authentication-operator failed liveness probe, will be restarted" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:12.767961 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" podUID="c3400b7c-5666-4b36-a50f-7800f6737527" containerName="authentication-operator" containerID="cri-o://ededff8a2f4ee4e72db8028a40598edd0ce0981b53e5b3d8e13b10053bf64cd1" gracePeriod=30 Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:12.815295 4824 patch_prober.go:28] interesting pod/apiserver-76f77b778f-pd82d container/openshift-apiserver namespace/openshift-apiserver: Liveness probe status=failure output="Get \"https://10.217.0.13:8443/livez?exclude=etcd\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:12.815360 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-apiserver/apiserver-76f77b778f-pd82d" podUID="7bff53ba-5406-445c-afcb-3247516a1258" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.13:8443/livez?exclude=etcd\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:12.815367 4824 patch_prober.go:28] interesting pod/apiserver-76f77b778f-pd82d container/openshift-apiserver namespace/openshift-apiserver: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz?exclude=etcd&exclude=etcd-readiness\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:12.815450 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-apiserver/apiserver-76f77b778f-pd82d" podUID="7bff53ba-5406-445c-afcb-3247516a1258" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz?exclude=etcd&exclude=etcd-readiness\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:12.914010 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8" podUID="23621a98-1d6b-421a-976c-965cecfe6db1" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.113:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:12.914103 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:12.996059 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8" podUID="23621a98-1d6b-421a-976c-965cecfe6db1" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.113:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:12.996108 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" podUID="d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:12.996193 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.160130 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" podUID="383a67da-0f83-41e2-82d6-809d1b6cdbec" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.118:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.160428 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" podUID="d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.324992 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl" podUID="6e91c44f-6221-4dfd-9c18-2bbedff02850" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.119:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.326220 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh" podUID="91bf6cd9-a4fe-46e8-b1e7-39e57c1398b3" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.408424 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" podUID="feab57e6-788b-4f23-9e99-aa248786052d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.408455 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh" podUID="91bf6cd9-a4fe-46e8-b1e7-39e57c1398b3" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.408688 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.577459 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="metrics-server" containerStatusID={"Type":"cri-o","ID":"7da26121a429a48f86bdeaadb323c671e73ba9993fa5bd40f3fcca78a5bf3769"} pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" containerMessage="Container metrics-server failed liveness probe, will be restarted" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.577527 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" podUID="e153e30f-4f5e-4a38-8efb-6452096e25d7" containerName="metrics-server" containerID="cri-o://7da26121a429a48f86bdeaadb323c671e73ba9993fa5bd40f3fcca78a5bf3769" gracePeriod=170 Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.623052 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" podUID="383a67da-0f83-41e2-82d6-809d1b6cdbec" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.118:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.623339 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.623495 4824 patch_prober.go:28] interesting pod/monitoring-plugin-6f74c7bb76-pbw94 container/monitoring-plugin namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.80:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.623548 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/monitoring-plugin-6f74c7bb76-pbw94" podUID="364b1858-6a46-4db4-9d9d-0ec19d54abc9" containerName="monitoring-plugin" probeResult="failure" output="Get \"https://10.217.0.80:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.624021 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl" podUID="6e91c44f-6221-4dfd-9c18-2bbedff02850" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.119:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.623101 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-q7psb" podUID="0af0249c-ed71-4f92-b1a3-cb6f43b92529" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.122:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.624679 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/monitoring-plugin-6f74c7bb76-pbw94" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.625180 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.706066 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9" podUID="31749799-9eb4-403f-a61d-9d50d8bc8367" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.117:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.788983 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9" podUID="31749799-9eb4-403f-a61d-9d50d8bc8367" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.117:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.789126 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.789155 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" podUID="feab57e6-788b-4f23-9e99-aa248786052d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.788994 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-5xfv2" podUID="87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.789285 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.789826 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4" podUID="143e48cd-c956-448c-8dcb-5858a582e29c" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.104:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.789882 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" podUID="bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.121:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.789921 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" podUID="bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.121:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.789972 4824 patch_prober.go:28] interesting pod/logging-loki-distributor-76cc67bf56-nl2vt container/loki-distributor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.51:3101/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.790007 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" podUID="72ff7b94-63fd-45d1-a803-bc1ccf4388ed" containerName="loki-distributor" probeResult="failure" output="Get \"https://10.217.0.51:3101/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.790346 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-q7psb" podUID="0af0249c-ed71-4f92-b1a3-cb6f43b92529" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.122:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.790521 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-5xfv2" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.790577 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.790616 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" Dec 09 11:14:13 crc kubenswrapper[4824]: I1209 11:14:13.831304 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6mfg8" podUID="5b34da01-28d8-476c-a2a5-a098489d90c1" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.102:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:13.831716 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" podUID="9f87669d-33a9-4269-808c-ca6c718f762c" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.44:6080/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:13.831840 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:13.872979 4824 patch_prober.go:28] interesting pod/console-operator-58897d9998-wgtx2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:13.873045 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" podUID="cd6c14a0-99d8-485d-81f3-cc4ebe5a943b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:13.873079 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-967d97867-lfprd" podUID="b0c20e13-b233-4b21-8f47-facef9603735" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:13.873119 4824 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-j4zxt container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:13.873249 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" podUID="7b26f121-04f6-4501-84b2-1833b927aa14" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:13.873312 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:13.873478 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" podUID="9f87669d-33a9-4269-808c-ca6c718f762c" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.44:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:13.873160 4824 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-j4zxt container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:13.873729 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" podUID="7b26f121-04f6-4501-84b2-1833b927aa14" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:13.873854 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:13.914476 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl" podUID="b6577050-6a21-43bb-84aa-20aad247aafc" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:13.914902 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-5xfv2" podUID="87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:13.914975 4824 patch_prober.go:28] interesting pod/logging-loki-querier-5895d59bb8-89dgf container/loki-querier namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.52:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:13.915007 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" podUID="1006b7c9-7244-44af-8bc9-52787d891f7f" containerName="loki-querier" probeResult="failure" output="Get \"https://10.217.0.52:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.000010 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-b9bnh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.000060 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b9bnh" podUID="6efb8245-07e9-4c31-b8ab-53ab3685e593" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.013558 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="metallb-system/speaker-5xfv2" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.013639 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.013650 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cert-manager-webhook" containerStatusID={"Type":"cri-o","ID":"1d23dc9a9237a313ad67f6e0183bda8cbae96cfe627b77a3827352d7702bd4b6"} pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" containerMessage="Container cert-manager-webhook failed liveness probe, will be restarted" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.013729 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" podUID="9f87669d-33a9-4269-808c-ca6c718f762c" containerName="cert-manager-webhook" containerID="cri-o://1d23dc9a9237a313ad67f6e0183bda8cbae96cfe627b77a3827352d7702bd4b6" gracePeriod=30 Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.013731 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="openshift-config-operator" containerStatusID={"Type":"cri-o","ID":"3efabae52732fd1dd695c8ec9ef93d6237ea498a410bac3039995980f09cd0d7"} pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" containerMessage="Container openshift-config-operator failed liveness probe, will be restarted" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.013871 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" podUID="7b26f121-04f6-4501-84b2-1833b927aa14" containerName="openshift-config-operator" containerID="cri-o://3efabae52732fd1dd695c8ec9ef93d6237ea498a410bac3039995980f09cd0d7" gracePeriod=30 Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.082097 4824 patch_prober.go:28] interesting pod/router-default-5444994796-mj7jx container/router namespace/openshift-ingress: Liveness probe status=failure output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.082156 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.082202 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.084093 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="router" containerStatusID={"Type":"cri-o","ID":"b81f987a695aac3b20392984022ac9b8c31c087c6d4439490168cb2c88e811c7"} pod="openshift-ingress/router-default-5444994796-mj7jx" containerMessage="Container router failed liveness probe, will be restarted" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.084141 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" containerID="cri-o://b81f987a695aac3b20392984022ac9b8c31c087c6d4439490168cb2c88e811c7" gracePeriod=10 Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.125238 4824 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-tg5lv container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.125310 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" podUID="45732d49-ce14-439b-8b7a-fb9fdf267fb1" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.125404 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.125441 4824 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-tg5lv container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.125510 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-b9bnh container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.125543 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-b9bnh" podUID="6efb8245-07e9-4c31-b8ab-53ab3685e593" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.125504 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" podUID="45732d49-ce14-439b-8b7a-fb9fdf267fb1" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.125612 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.207968 4824 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-9c2mr container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.208024 4824 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-9c2mr container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.208039 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" podUID="62823972-71dc-4f7d-b4f6-da0cb90dcf36" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.208076 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" podUID="62823972-71dc-4f7d-b4f6-da0cb90dcf36" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.208025 4824 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-k8bgl container/package-server-manager namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.208104 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.208123 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" podUID="6c10edf9-cf75-4633-8d4c-9bbb9d9b6339" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.208146 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.208198 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.249174 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" podUID="d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.249338 4824 patch_prober.go:28] interesting pod/router-default-5444994796-mj7jx container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.249579 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.249707 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.249763 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8" podUID="23621a98-1d6b-421a-976c-965cecfe6db1" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.113:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.249844 4824 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-k8bgl container/package-server-manager namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.249873 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" podUID="6c10edf9-cf75-4633-8d4c-9bbb9d9b6339" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.249936 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.282543 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": context deadline exceeded" start-of-body= Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.282607 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": context deadline exceeded" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.307808 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:6443/livez?exclude=etcd\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.307864 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez?exclude=etcd\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.429316 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="package-server-manager" containerStatusID={"Type":"cri-o","ID":"886bc9d74183124b07bb450444597fe02f70436ac80f07efff6ededc8ff84914"} pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" containerMessage="Container package-server-manager failed liveness probe, will be restarted" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.429379 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" podUID="6c10edf9-cf75-4633-8d4c-9bbb9d9b6339" containerName="package-server-manager" containerID="cri-o://886bc9d74183124b07bb450444597fe02f70436ac80f07efff6ededc8ff84914" gracePeriod=30 Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.441797 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="catalog-operator" containerStatusID={"Type":"cri-o","ID":"e048faa53b6a634fed8d590a6500c0a2619064882a9070e599eea82d1a432a34"} pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" containerMessage="Container catalog-operator failed liveness probe, will be restarted" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.441825 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="packageserver" containerStatusID={"Type":"cri-o","ID":"8424dba8744dda5f7cdf5e3a63688e51799cdbdb07d370b2b3b2811ce0148354"} pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" containerMessage="Container packageserver failed liveness probe, will be restarted" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.441873 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" podUID="62823972-71dc-4f7d-b4f6-da0cb90dcf36" containerName="catalog-operator" containerID="cri-o://e048faa53b6a634fed8d590a6500c0a2619064882a9070e599eea82d1a432a34" gracePeriod=30 Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.441883 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" podUID="45732d49-ce14-439b-8b7a-fb9fdf267fb1" containerName="packageserver" containerID="cri-o://8424dba8744dda5f7cdf5e3a63688e51799cdbdb07d370b2b3b2811ce0148354" gracePeriod=30 Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.452011 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh" podUID="91bf6cd9-a4fe-46e8-b1e7-39e57c1398b3" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.554211 4824 patch_prober.go:28] interesting pod/logging-loki-distributor-76cc67bf56-nl2vt container/loki-distributor namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.51:3101/loki/api/v1/status/buildinfo\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.554667 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" podUID="72ff7b94-63fd-45d1-a803-bc1ccf4388ed" containerName="loki-distributor" probeResult="failure" output="Get \"https://10.217.0.51:3101/loki/api/v1/status/buildinfo\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.708278 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl" podUID="6e91c44f-6221-4dfd-9c18-2bbedff02850" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.119:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:14 crc kubenswrapper[4824]: I1209 11:14:14.708301 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" podUID="383a67da-0f83-41e2-82d6-809d1b6cdbec" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.118:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.010275 4824 patch_prober.go:28] interesting pod/logging-loki-query-frontend-84558f7c9f-9d5fm container/loki-query-frontend namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.010359 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" podUID="e05c11bd-cd92-4bac-adea-0a6049ccfb39" containerName="loki-query-frontend" probeResult="failure" output="Get \"https://10.217.0.53:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.010473 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.011217 4824 patch_prober.go:28] interesting pod/monitoring-plugin-6f74c7bb76-pbw94 container/monitoring-plugin namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.80:9443/health\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.011252 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/monitoring-plugin-6f74c7bb76-pbw94" podUID="364b1858-6a46-4db4-9d9d-0ec19d54abc9" containerName="monitoring-plugin" probeResult="failure" output="Get \"https://10.217.0.80:9443/health\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.011280 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-vz44w container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.011303 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" podUID="93c989b8-7b86-4339-bbab-5886c7d13dc9" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.55:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.012250 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-bvp6v container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.012314 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" podUID="858d899b-800f-4639-8fbd-4f1ccad44991" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.012346 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-vz44w container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:8083/ready\": context deadline exceeded" start-of-body= Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.012361 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" podUID="93c989b8-7b86-4339-bbab-5886c7d13dc9" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.55:8083/ready\": context deadline exceeded" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.051986 4824 patch_prober.go:28] interesting pod/logging-loki-querier-5895d59bb8-89dgf container/loki-querier namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.52:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.052058 4824 patch_prober.go:28] interesting pod/console-7c85c585b4-tgrff container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.139:8443/health\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.052099 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" podUID="1006b7c9-7244-44af-8bc9-52787d891f7f" containerName="loki-querier" probeResult="failure" output="Get \"https://10.217.0.52:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.052136 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-7c85c585b4-tgrff" podUID="e61278e7-d0a8-4039-909d-c2812c8a4a81" containerName="console" probeResult="failure" output="Get \"https://10.217.0.139:8443/health\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.052253 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.084799 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="speaker" containerStatusID={"Type":"cri-o","ID":"dd1807b856b18c97d0d6ae44f7011055c8dcd704fa540f6f20146e7f61f0b645"} pod="metallb-system/speaker-5xfv2" containerMessage="Container speaker failed liveness probe, will be restarted" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.084890 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/speaker-5xfv2" podUID="87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe" containerName="speaker" containerID="cri-o://dd1807b856b18c97d0d6ae44f7011055c8dcd704fa540f6f20146e7f61f0b645" gracePeriod=2 Dec 09 11:14:15 crc kubenswrapper[4824]: E1209 11:14:15.105827 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a13a50a9012203c512c34fec894886dfe2d054768904caa6f9ab13fa1fc56c87" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 09 11:14:15 crc kubenswrapper[4824]: E1209 11:14:15.107252 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a13a50a9012203c512c34fec894886dfe2d054768904caa6f9ab13fa1fc56c87" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 09 11:14:15 crc kubenswrapper[4824]: E1209 11:14:15.109202 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a13a50a9012203c512c34fec894886dfe2d054768904caa6f9ab13fa1fc56c87" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 09 11:14:15 crc kubenswrapper[4824]: E1209 11:14:15.109248 4824 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="c2dd0680-d32f-45e8-b834-6a327d2eaa21" containerName="ovn-northd" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.218949 4824 patch_prober.go:28] interesting pod/logging-loki-distributor-76cc67bf56-nl2vt container/loki-distributor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.51:3101/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.218994 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" podUID="bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.121:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.219029 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" podUID="72ff7b94-63fd-45d1-a803-bc1ccf4388ed" containerName="loki-distributor" probeResult="failure" output="Get \"https://10.217.0.51:3101/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.257384 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-bvp6v container/opa namespace/openshift-logging: Liveness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body={ Dec 09 11:14:15 crc kubenswrapper[4824]: "http": "Get \"http://localhost:8082\": context deadline exceeded" Dec 09 11:14:15 crc kubenswrapper[4824]: } Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.257454 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" podUID="858d899b-800f-4639-8fbd-4f1ccad44991" containerName="opa" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.259929 4824 patch_prober.go:28] interesting pod/loki-operator-controller-manager-6bf69c47b7-8fjw7 container/manager namespace/openshift-operators-redhat: Readiness probe status=failure output="Get \"http://10.217.0.47:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.259970 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" podUID="352c7865-5a0b-4ebb-93f2-513ce433ff8f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.47:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.259995 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-bvp6v container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8081/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.260079 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" podUID="858d899b-800f-4639-8fbd-4f1ccad44991" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.54:8081/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.260135 4824 patch_prober.go:28] interesting pod/logging-loki-querier-5895d59bb8-89dgf container/loki-querier namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.52:3101/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.260161 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" podUID="1006b7c9-7244-44af-8bc9-52787d891f7f" containerName="loki-querier" probeResult="failure" output="Get \"https://10.217.0.52:3101/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.260445 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9" podUID="31749799-9eb4-403f-a61d-9d50d8bc8367" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.117:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.260489 4824 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-tg5lv container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.260505 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" podUID="45732d49-ce14-439b-8b7a-fb9fdf267fb1" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.260532 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" podUID="feab57e6-788b-4f23-9e99-aa248786052d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.260557 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-5xfv2" podUID="87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.260738 4824 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-9c2mr container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.260799 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" podUID="62823972-71dc-4f7d-b4f6-da0cb90dcf36" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.260822 4824 patch_prober.go:28] interesting pod/oauth-openshift-f578d5c8f-5jzzv container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.62:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.260892 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" podUID="3d1e2cd1-5fc5-4a76-b1ff-07db11b07708" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.62:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.260972 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.260999 4824 patch_prober.go:28] interesting pod/oauth-openshift-f578d5c8f-5jzzv container/oauth-openshift namespace/openshift-authentication: Liveness probe status=failure output="Get \"https://10.217.0.62:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.261024 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" podUID="3d1e2cd1-5fc5-4a76-b1ff-07db11b07708" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.62:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.261057 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.262877 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="oauth-openshift" containerStatusID={"Type":"cri-o","ID":"166928d0068d6f77eb364346c00c3dd78dbb7602e7badaa6af86f3ce7f5c0df1"} pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" containerMessage="Container oauth-openshift failed liveness probe, will be restarted" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.306595 4824 patch_prober.go:28] interesting pod/logging-loki-compactor-0 container/loki-compactor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.65:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.306649 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-compactor-0" podUID="379b9136-1cdf-4786-8494-e99c7161b202" containerName="loki-compactor" probeResult="failure" output="Get \"https://10.217.0.65:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.306728 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-compactor-0" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.374325 4824 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.64:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.374403 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="9d24f296-729f-4eec-a1ae-d6b904399394" containerName="loki-ingester" probeResult="failure" output="Get \"https://10.217.0.64:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.374512 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-ingester-0" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.386861 4824 patch_prober.go:28] interesting pod/image-registry-66df7c8f76-jq6bp container/registry namespace/openshift-image-registry: Readiness probe status=failure output="Get \"https://10.217.0.73:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.386922 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" podUID="ffa85659-bcb5-43c8-9270-ab07d6ec9b70" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.73:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.387096 4824 patch_prober.go:28] interesting pod/image-registry-66df7c8f76-jq6bp container/registry namespace/openshift-image-registry: Liveness probe status=failure output="Get \"https://10.217.0.73:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.387149 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-image-registry/image-registry-66df7c8f76-jq6bp" podUID="ffa85659-bcb5-43c8-9270-ab07d6ec9b70" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.73:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.455632 4824 patch_prober.go:28] interesting pod/logging-loki-index-gateway-0 container/loki-index-gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.74:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.455728 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-index-gateway-0" podUID="6baf0bfe-7be6-4309-93da-e86174e4654e" containerName="loki-index-gateway" probeResult="failure" output="Get \"https://10.217.0.74:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.455852 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.559754 4824 patch_prober.go:28] interesting pod/logging-loki-query-frontend-84558f7c9f-9d5fm container/loki-query-frontend namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.53:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.559827 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" podUID="e05c11bd-cd92-4bac-adea-0a6049ccfb39" containerName="loki-query-frontend" probeResult="failure" output="Get \"https://10.217.0.53:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.679542 4824 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-j4zxt container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.679596 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" podUID="7b26f121-04f6-4501-84b2-1833b927aa14" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.688860 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-vz44w container/gateway namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.55:8081/live\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.688928 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" podUID="93c989b8-7b86-4339-bbab-5886c7d13dc9" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.55:8081/live\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.802107 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-controller-ovs-rl4nw" podUID="2a54d12b-327d-409e-8652-9525878ae96f" containerName="ovs-vswitchd" probeResult="failure" output="command timed out" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.802155 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-ovs-rl4nw" podUID="2a54d12b-327d-409e-8652-9525878ae96f" containerName="ovs-vswitchd" probeResult="failure" output="command timed out" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.802324 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-index-2jnrg" podUID="25c6995d-6a34-43fc-9b1b-6c34ed9aec46" containerName="registry-server" probeResult="failure" output="command timed out" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.802372 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-controller-ovs-rl4nw" podUID="2a54d12b-327d-409e-8652-9525878ae96f" containerName="ovsdb-server" probeResult="failure" output="command timed out" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.802375 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-ovs-rl4nw" podUID="2a54d12b-327d-409e-8652-9525878ae96f" containerName="ovsdb-server" probeResult="failure" output="command timed out" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.802455 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-index-2jnrg" podUID="25c6995d-6a34-43fc-9b1b-6c34ed9aec46" containerName="registry-server" probeResult="failure" output="command timed out" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.825161 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7c85c585b4-tgrff" Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.851568 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-bvp6v container/gateway namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.54:8081/live\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:15 crc kubenswrapper[4824]: I1209 11:14:15.852036 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" podUID="858d899b-800f-4639-8fbd-4f1ccad44991" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.54:8081/live\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:15 crc kubenswrapper[4824]: E1209 11:14:15.877215 4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:16 crc kubenswrapper[4824]: I1209 11:14:16.050180 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" podUID="e7e496dc-23c4-47cd-9c5a-aa4430130849" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.123:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:16 crc kubenswrapper[4824]: I1209 11:14:16.050301 4824 patch_prober.go:28] interesting pod/logging-loki-query-frontend-84558f7c9f-9d5fm container/loki-query-frontend namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:16 crc kubenswrapper[4824]: I1209 11:14:16.050349 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 11:14:16 crc kubenswrapper[4824]: I1209 11:14:16.050382 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" podUID="e05c11bd-cd92-4bac-adea-0a6049ccfb39" containerName="loki-query-frontend" probeResult="failure" output="Get \"https://10.217.0.53:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:16 crc kubenswrapper[4824]: I1209 11:14:16.301964 4824 patch_prober.go:28] interesting pod/oauth-openshift-f578d5c8f-5jzzv container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.62:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:16 crc kubenswrapper[4824]: I1209 11:14:16.302203 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" podUID="3d1e2cd1-5fc5-4a76-b1ff-07db11b07708" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.62:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:16 crc kubenswrapper[4824]: I1209 11:14:16.302019 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-5xfv2" podUID="87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:16 crc kubenswrapper[4824]: I1209 11:14:16.306409 4824 patch_prober.go:28] interesting pod/logging-loki-compactor-0 container/loki-compactor namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.65:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:16 crc kubenswrapper[4824]: I1209 11:14:16.306493 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-compactor-0" podUID="379b9136-1cdf-4786-8494-e99c7161b202" containerName="loki-compactor" probeResult="failure" output="Get \"https://10.217.0.65:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:16 crc kubenswrapper[4824]: I1209 11:14:16.307511 4824 patch_prober.go:28] interesting pod/logging-loki-compactor-0 container/loki-compactor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.65:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:16 crc kubenswrapper[4824]: I1209 11:14:16.307624 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-compactor-0" podUID="379b9136-1cdf-4786-8494-e99c7161b202" containerName="loki-compactor" probeResult="failure" output="Get \"https://10.217.0.65:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:16 crc kubenswrapper[4824]: I1209 11:14:16.374118 4824 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.64:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:16 crc kubenswrapper[4824]: I1209 11:14:16.374224 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-ingester-0" podUID="9d24f296-729f-4eec-a1ae-d6b904399394" containerName="loki-ingester" probeResult="failure" output="Get \"https://10.217.0.64:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:16 crc kubenswrapper[4824]: I1209 11:14:16.375938 4824 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.64:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:16 crc kubenswrapper[4824]: I1209 11:14:16.376038 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="9d24f296-729f-4eec-a1ae-d6b904399394" containerName="loki-ingester" probeResult="failure" output="Get \"https://10.217.0.64:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:16 crc kubenswrapper[4824]: I1209 11:14:16.455269 4824 patch_prober.go:28] interesting pod/logging-loki-index-gateway-0 container/loki-index-gateway namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.74:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:16 crc kubenswrapper[4824]: I1209 11:14:16.455342 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-index-gateway-0" podUID="6baf0bfe-7be6-4309-93da-e86174e4654e" containerName="loki-index-gateway" probeResult="failure" output="Get \"https://10.217.0.74:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:16 crc kubenswrapper[4824]: I1209 11:14:16.457344 4824 patch_prober.go:28] interesting pod/logging-loki-index-gateway-0 container/loki-index-gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.74:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:16 crc kubenswrapper[4824]: I1209 11:14:16.457411 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-index-gateway-0" podUID="6baf0bfe-7be6-4309-93da-e86174e4654e" containerName="loki-index-gateway" probeResult="failure" output="Get \"https://10.217.0.74:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:16 crc kubenswrapper[4824]: I1209 11:14:16.745246 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" podUID="34793d58-1049-48ff-b8e7-187bdf42e550" containerName="sbdb" probeResult="failure" output="command timed out" Dec 09 11:14:16 crc kubenswrapper[4824]: I1209 11:14:16.761329 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-2qqhd" podUID="34793d58-1049-48ff-b8e7-187bdf42e550" containerName="nbdb" probeResult="failure" output="command timed out" Dec 09 11:14:17 crc kubenswrapper[4824]: I1209 11:14:17.095036 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" podUID="e7e496dc-23c4-47cd-9c5a-aa4430130849" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.123:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:17 crc kubenswrapper[4824]: I1209 11:14:17.095098 4824 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:17 crc kubenswrapper[4824]: I1209 11:14:17.095139 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:17 crc kubenswrapper[4824]: I1209 11:14:17.095214 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="54621281-db00-41e4-b617-032435893391" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.0.254:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:17 crc kubenswrapper[4824]: I1209 11:14:17.095231 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 11:14:17 crc kubenswrapper[4824]: I1209 11:14:17.095278 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 09 11:14:17 crc kubenswrapper[4824]: I1209 11:14:17.095349 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/kube-state-metrics-0" podUID="54621281-db00-41e4-b617-032435893391" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.0.254:8080/livez\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:17 crc kubenswrapper[4824]: I1209 11:14:17.095500 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/kube-state-metrics-0" Dec 09 11:14:17 crc kubenswrapper[4824]: I1209 11:14:17.097086 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-state-metrics" containerStatusID={"Type":"cri-o","ID":"2bbad6105a2324d3ba9b939f829736af090521c9ba0ff0e211e1344cab369da7"} pod="openstack/kube-state-metrics-0" containerMessage="Container kube-state-metrics failed liveness probe, will be restarted" Dec 09 11:14:17 crc kubenswrapper[4824]: I1209 11:14:17.097290 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="54621281-db00-41e4-b617-032435893391" containerName="kube-state-metrics" containerID="cri-o://2bbad6105a2324d3ba9b939f829736af090521c9ba0ff0e211e1344cab369da7" gracePeriod=30 Dec 09 11:14:17 crc kubenswrapper[4824]: I1209 11:14:17.110417 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/prometheus-metric-storage-0" podUID="fb8f1e44-4c21-423a-bacc-1cca7d7715c3" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.167:9090/-/healthy\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:17 crc kubenswrapper[4824]: I1209 11:14:17.110498 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="fb8f1e44-4c21-423a-bacc-1cca7d7715c3" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.167:9090/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:17 crc kubenswrapper[4824]: I1209 11:14:17.124114 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="54621281-db00-41e4-b617-032435893391" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.0.254:8081/readyz\": read tcp 10.217.0.2:37212->10.217.0.254:8081: read: connection reset by peer" Dec 09 11:14:17 crc kubenswrapper[4824]: I1209 11:14:17.233414 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" podUID="9f87669d-33a9-4269-808c-ca6c718f762c" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.44:6080/healthz\": dial tcp 10.217.0.44:6080: connect: connection refused" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:17.743363 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-handler-5strj" podUID="952171d0-1a2e-4801-bc11-4012ba19588c" containerName="nmstate-handler" probeResult="failure" output="command timed out" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:17.938322 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="153fe73c-2767-4f80-a7b4-0af49257065a" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.208:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.022097 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" podUID="1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.108:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.022270 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.096829 4824 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.096885 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.161808 4824 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-n8zgd container/oauth-apiserver namespace/openshift-oauth-apiserver: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.161893 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n8zgd" podUID="f93c02c2-b12a-489d-9f82-1ca3dadd18d5" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.22:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.343980 4824 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-4dlng container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.344042 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" podUID="86c42164-acb8-4a27-9a0f-48307cd304e4" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.344164 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.344435 4824 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-4dlng container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.344518 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" podUID="86c42164-acb8-4a27-9a0f-48307cd304e4" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.344591 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.590918 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" podUID="c2889de3-49b2-4465-8235-37f9e58b42a3" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.116:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.591036 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.680307 4824 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-j4zxt container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.680415 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" podUID="7b26f121-04f6-4501-84b2-1833b927aa14" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.688494 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="prometheus-operator-admission-webhook" containerStatusID={"Type":"cri-o","ID":"00bae8ec6f6e4cf5dd4670db60b6420c115bdac303b2c253cc78e718e60f743c"} pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" containerMessage="Container prometheus-operator-admission-webhook failed liveness probe, will be restarted" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.688572 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" podUID="86c42164-acb8-4a27-9a0f-48307cd304e4" containerName="prometheus-operator-admission-webhook" containerID="cri-o://00bae8ec6f6e4cf5dd4670db60b6420c115bdac303b2c253cc78e718e60f743c" gracePeriod=30 Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.727983 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.761281 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="56148e0f-636f-410e-bfb5-342da01e8c76" containerName="galera" probeResult="failure" output="command timed out" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.761471 4824 generic.go:334] "Generic (PLEG): container finished" podID="9f87669d-33a9-4269-808c-ca6c718f762c" containerID="1d23dc9a9237a313ad67f6e0183bda8cbae96cfe627b77a3827352d7702bd4b6" exitCode=0 Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.761863 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="844884cd-6813-4fa2-88e4-7a5994bbd3cb" containerName="prometheus" probeResult="failure" output="command timed out" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.762200 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-k8s-0" podUID="844884cd-6813-4fa2-88e4-7a5994bbd3cb" containerName="prometheus" probeResult="failure" output="command timed out" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.763447 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" event={"ID":"9f87669d-33a9-4269-808c-ca6c718f762c","Type":"ContainerDied","Data":"1d23dc9a9237a313ad67f6e0183bda8cbae96cfe627b77a3827352d7702bd4b6"} Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.830576 4824 generic.go:334] "Generic (PLEG): container finished" podID="62823972-71dc-4f7d-b4f6-da0cb90dcf36" containerID="e048faa53b6a634fed8d590a6500c0a2619064882a9070e599eea82d1a432a34" exitCode=0 Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.830652 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" event={"ID":"62823972-71dc-4f7d-b4f6-da0cb90dcf36","Type":"ContainerDied","Data":"e048faa53b6a634fed8d590a6500c0a2619064882a9070e599eea82d1a432a34"} Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.834863 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c2dd0680-d32f-45e8-b834-6a327d2eaa21/ovn-northd/0.log" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.834969 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c2dd0680-d32f-45e8-b834-6a327d2eaa21","Type":"ContainerDied","Data":"a13a50a9012203c512c34fec894886dfe2d054768904caa6f9ab13fa1fc56c87"} Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.834898 4824 generic.go:334] "Generic (PLEG): container finished" podID="c2dd0680-d32f-45e8-b834-6a327d2eaa21" containerID="a13a50a9012203c512c34fec894886dfe2d054768904caa6f9ab13fa1fc56c87" exitCode=139 Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.837479 4824 generic.go:334] "Generic (PLEG): container finished" podID="87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe" containerID="dd1807b856b18c97d0d6ae44f7011055c8dcd704fa540f6f20146e7f61f0b645" exitCode=137 Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:18.837502 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5xfv2" event={"ID":"87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe","Type":"ContainerDied","Data":"dd1807b856b18c97d0d6ae44f7011055c8dcd704fa540f6f20146e7f61f0b645"} Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.064034 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" podUID="1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.108:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.305039 4824 patch_prober.go:28] interesting pod/route-controller-manager-7f84c5b8f9-rqmr9 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.305084 4824 patch_prober.go:28] interesting pod/route-controller-manager-7f84c5b8f9-rqmr9 container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.305101 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" podUID="afe92d2b-bce3-445e-9c26-9c533353e7a3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.305116 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" podUID="afe92d2b-bce3-445e-9c26-9c533353e7a3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.305151 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.306511 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="route-controller-manager" containerStatusID={"Type":"cri-o","ID":"03c338c0a6a98a41c144c02fffabcfe6e5a3daf642e4e2523e58122af8753be4"} pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" containerMessage="Container route-controller-manager failed liveness probe, will be restarted" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.306549 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" podUID="afe92d2b-bce3-445e-9c26-9c533353e7a3" containerName="route-controller-manager" containerID="cri-o://03c338c0a6a98a41c144c02fffabcfe6e5a3daf642e4e2523e58122af8753be4" gracePeriod=30 Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.458885 4824 trace.go:236] Trace[2007644081]: "Calculate volume metrics of wal for pod openshift-logging/logging-loki-ingester-0" (09-Dec-2025 11:14:10.062) (total time: 9392ms): Dec 09 11:14:21 crc kubenswrapper[4824]: Trace[2007644081]: [9.392685481s] [9.392685481s] END Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.458887 4824 trace.go:236] Trace[590391067]: "Calculate volume metrics of ovndbcluster-nb-etc-ovn for pod openstack/ovsdbserver-nb-0" (09-Dec-2025 11:14:12.503) (total time: 6953ms): Dec 09 11:14:21 crc kubenswrapper[4824]: Trace[590391067]: [6.953037738s] [6.953037738s] END Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.458932 4824 trace.go:236] Trace[1049360355]: "Calculate volume metrics of persistence for pod openstack/rabbitmq-server-2" (09-Dec-2025 11:14:08.962) (total time: 10496ms): Dec 09 11:14:21 crc kubenswrapper[4824]: Trace[1049360355]: [10.496536891s] [10.496536891s] END Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.460325 4824 trace.go:236] Trace[1856239280]: "Calculate volume metrics of glance for pod openstack/glance-default-external-api-0" (09-Dec-2025 11:14:17.319) (total time: 2141ms): Dec 09 11:14:21 crc kubenswrapper[4824]: Trace[1856239280]: [2.141197367s] [2.141197367s] END Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.458930 4824 trace.go:236] Trace[2044589940]: "Calculate volume metrics of storage for pod openshift-logging/logging-loki-compactor-0" (09-Dec-2025 11:14:09.191) (total time: 10267ms): Dec 09 11:14:21 crc kubenswrapper[4824]: Trace[2044589940]: [10.267790292s] [10.267790292s] END Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.574919 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.697135 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-vz44w container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.697423 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b549956cc-vz44w" podUID="93c989b8-7b86-4339-bbab-5886c7d13dc9" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.55:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.758365 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-95jfr" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" containerName="registry-server" probeResult="failure" output="command timed out" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.758426 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-operators-582xj" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" containerName="registry-server" probeResult="failure" output="command timed out" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.758501 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-95jfr" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.758540 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-582xj" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.758539 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-marketplace-95jfr" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" containerName="registry-server" probeResult="failure" output="command timed out" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.758568 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-operators-582xj" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" containerName="registry-server" probeResult="failure" output="command timed out" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.758634 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-95jfr" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.758663 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/redhat-operators-582xj" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.761323 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="registry-server" containerStatusID={"Type":"cri-o","ID":"b70d423ba00ad8d15b6e60bc0af43894068b4f8a5957df3d2fa34df475dfb2cf"} pod="openshift-marketplace/redhat-operators-582xj" containerMessage="Container registry-server failed liveness probe, will be restarted" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.761371 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-582xj" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" containerName="registry-server" containerID="cri-o://b70d423ba00ad8d15b6e60bc0af43894068b4f8a5957df3d2fa34df475dfb2cf" gracePeriod=30 Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.761327 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="registry-server" containerStatusID={"Type":"cri-o","ID":"1cf855518b8661ca5358ae1fd9e5a18c8e9fddda7448fb2fd36686a961879490"} pod="openshift-marketplace/redhat-marketplace-95jfr" containerMessage="Container registry-server failed liveness probe, will be restarted" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.761420 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-95jfr" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" containerName="registry-server" containerID="cri-o://1cf855518b8661ca5358ae1fd9e5a18c8e9fddda7448fb2fd36686a961879490" gracePeriod=30 Dec 09 11:14:21 crc kubenswrapper[4824]: E1209 11:14:19.771568 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1cf855518b8661ca5358ae1fd9e5a18c8e9fddda7448fb2fd36686a961879490" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 11:14:21 crc kubenswrapper[4824]: E1209 11:14:19.771572 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b70d423ba00ad8d15b6e60bc0af43894068b4f8a5957df3d2fa34df475dfb2cf" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 11:14:21 crc kubenswrapper[4824]: E1209 11:14:19.773676 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1cf855518b8661ca5358ae1fd9e5a18c8e9fddda7448fb2fd36686a961879490" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 11:14:21 crc kubenswrapper[4824]: E1209 11:14:19.773975 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b70d423ba00ad8d15b6e60bc0af43894068b4f8a5957df3d2fa34df475dfb2cf" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 11:14:21 crc kubenswrapper[4824]: E1209 11:14:19.775311 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1cf855518b8661ca5358ae1fd9e5a18c8e9fddda7448fb2fd36686a961879490" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 11:14:21 crc kubenswrapper[4824]: E1209 11:14:19.775371 4824 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-95jfr" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" containerName="registry-server" Dec 09 11:14:21 crc kubenswrapper[4824]: E1209 11:14:19.775646 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b70d423ba00ad8d15b6e60bc0af43894068b4f8a5957df3d2fa34df475dfb2cf" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 11:14:21 crc kubenswrapper[4824]: E1209 11:14:19.775689 4824 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-marketplace/redhat-operators-582xj" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" containerName="registry-server" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.851096 4824 patch_prober.go:28] interesting pod/logging-loki-gateway-b549956cc-bvp6v container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.851164 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b549956cc-bvp6v" podUID="858d899b-800f-4639-8fbd-4f1ccad44991" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.54:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.976039 4824 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:19.976132 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.195954 4824 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Liveness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.196113 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.196169 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.197896 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-scheduler" containerStatusID={"Type":"cri-o","ID":"e1e6cfeb997cafb7ec7f6d829e1ff2e2d95e4b343c422194a5290d222ea53a7a"} pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" containerMessage="Container kube-scheduler failed liveness probe, will be restarted" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.197990 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" containerID="cri-o://e1e6cfeb997cafb7ec7f6d829e1ff2e2d95e4b343c422194a5290d222ea53a7a" gracePeriod=30 Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.433615 4824 generic.go:334] "Generic (PLEG): container finished" podID="54621281-db00-41e4-b617-032435893391" containerID="2bbad6105a2324d3ba9b939f829736af090521c9ba0ff0e211e1344cab369da7" exitCode=2 Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.433666 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"54621281-db00-41e4-b617-032435893391","Type":"ContainerDied","Data":"2bbad6105a2324d3ba9b939f829736af090521c9ba0ff0e211e1344cab369da7"} Dec 09 11:14:21 crc kubenswrapper[4824]: E1209 11:14:20.453242 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a13a50a9012203c512c34fec894886dfe2d054768904caa6f9ab13fa1fc56c87 is running failed: container process not found" containerID="a13a50a9012203c512c34fec894886dfe2d054768904caa6f9ab13fa1fc56c87" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 09 11:14:21 crc kubenswrapper[4824]: E1209 11:14:20.455481 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a13a50a9012203c512c34fec894886dfe2d054768904caa6f9ab13fa1fc56c87 is running failed: container process not found" containerID="a13a50a9012203c512c34fec894886dfe2d054768904caa6f9ab13fa1fc56c87" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 09 11:14:21 crc kubenswrapper[4824]: E1209 11:14:20.461478 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a13a50a9012203c512c34fec894886dfe2d054768904caa6f9ab13fa1fc56c87 is running failed: container process not found" containerID="a13a50a9012203c512c34fec894886dfe2d054768904caa6f9ab13fa1fc56c87" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 09 11:14:21 crc kubenswrapper[4824]: E1209 11:14:20.461550 4824 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a13a50a9012203c512c34fec894886dfe2d054768904caa6f9ab13fa1fc56c87 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="c2dd0680-d32f-45e8-b834-6a327d2eaa21" containerName="ovn-northd" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.470964 4824 patch_prober.go:28] interesting pod/thanos-querier-7fc6d7f97-nsq7f container/kube-rbac-proxy-web namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.72:9091/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.471008 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-7fc6d7f97-nsq7f" podUID="d030e6a9-d47a-47bf-9c24-ff0ca58f71e6" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.72:9091/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.471178 4824 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-qn2qz container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.5:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.471258 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" podUID="612ebf47-a6c4-4a62-8d97-91a003f49c44" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.5:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.471359 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.471520 4824 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-qn2qz container/operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.5:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.472098 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" podUID="612ebf47-a6c4-4a62-8d97-91a003f49c44" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.5:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.472188 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.495604 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="operator" containerStatusID={"Type":"cri-o","ID":"6211e2b6e456aecff8e6484685cb1dbfb8ed11c2a63a1ec44ca688b1a82a5c22"} pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" containerMessage="Container operator failed liveness probe, will be restarted" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.497499 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" podUID="612ebf47-a6c4-4a62-8d97-91a003f49c44" containerName="operator" containerID="cri-o://6211e2b6e456aecff8e6484685cb1dbfb8ed11c2a63a1ec44ca688b1a82a5c22" gracePeriod=30 Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.578657 4824 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.578874 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.744513 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="73d6bd70-44c7-4eed-a93a-36df636869cf" containerName="galera" probeResult="failure" output="command timed out" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.744607 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="73d6bd70-44c7-4eed-a93a-36df636869cf" containerName="galera" probeResult="failure" output="command timed out" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.744663 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.745920 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="galera" containerStatusID={"Type":"cri-o","ID":"3187d927bee90adbfb095796189ae710aa8a824436cd0f34583d4da240bacb9d"} pod="openstack/openstack-cell1-galera-0" containerMessage="Container galera failed liveness probe, will be restarted" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.747118 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-jm5ld" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" containerName="registry-server" probeResult="failure" output="command timed out" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.748289 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/certified-operators-2cbmd" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" containerName="registry-server" probeResult="failure" output="command timed out" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.748519 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/community-operators-jm5ld" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" containerName="registry-server" probeResult="failure" output="command timed out" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.749536 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/certified-operators-2cbmd" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" containerName="registry-server" probeResult="failure" output="command timed out" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.800847 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jm5ld" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.800980 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.801058 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2cbmd" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.801082 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/community-operators-jm5ld" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.801091 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/certified-operators-2cbmd" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.804111 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="registry-server" containerStatusID={"Type":"cri-o","ID":"4837611c2a862dc0e89e9c677ca72373bccca69d09dbaaa35de3699394b0a701"} pod="openshift-marketplace/community-operators-jm5ld" containerMessage="Container registry-server failed liveness probe, will be restarted" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.804227 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jm5ld" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" containerName="registry-server" containerID="cri-o://4837611c2a862dc0e89e9c677ca72373bccca69d09dbaaa35de3699394b0a701" gracePeriod=30 Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.805461 4824 patch_prober.go:28] interesting pod/perses-operator-5446b9c989-pkmss container/perses-operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.30:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.805559 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/perses-operator-5446b9c989-pkmss" podUID="2ecf9e90-37fe-416d-9e7a-400373cfbc8d" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.30:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.806116 4824 patch_prober.go:28] interesting pod/perses-operator-5446b9c989-pkmss container/perses-operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.30:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.806204 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/perses-operator-5446b9c989-pkmss" podUID="2ecf9e90-37fe-416d-9e7a-400373cfbc8d" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.30:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.806332 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-pkmss" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.813229 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="registry-server" containerStatusID={"Type":"cri-o","ID":"8eb26eb28020a6b331e841e6bc1a33e26e08169614eddc4b7de37f69670eabf9"} pod="openshift-marketplace/certified-operators-2cbmd" containerMessage="Container registry-server failed liveness probe, will be restarted" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.813331 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2cbmd" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" containerName="registry-server" containerID="cri-o://8eb26eb28020a6b331e841e6bc1a33e26e08169614eddc4b7de37f69670eabf9" gracePeriod=30 Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.848171 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" podUID="bab491e7-4eec-4ba1-975d-fb2468372ade" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.95:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.848483 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.890095 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" podUID="bab491e7-4eec-4ba1-975d-fb2468372ade" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.95:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:20.890379 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.001672 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-pkmss" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.008085 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.009556 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-6mfg8" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.129975 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4" podUID="143e48cd-c956-448c-8dcb-5858a582e29c" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.104:8081/readyz\": read tcp 10.217.0.2:60172->10.217.0.104:8081: read: connection reset by peer" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.216001 4824 patch_prober.go:28] interesting pod/controller-manager-69949c7bd4-pmmk2 container/controller-manager namespace/openshift-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.78:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.216342 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" podUID="ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.78:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.216404 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.216583 4824 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.216627 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.217089 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-967d97867-lfprd" podUID="b0c20e13-b233-4b21-8f47-facef9603735" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/readyz\": dial tcp 10.217.0.109:8081: connect: connection refused" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.216029 4824 patch_prober.go:28] interesting pod/controller-manager-69949c7bd4-pmmk2 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.78:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.217134 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" podUID="ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.78:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.217640 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="controller-manager" containerStatusID={"Type":"cri-o","ID":"8ef0f4d72f39e4c91f5d814ad0f5248f69fc7641a71cf06a32de5125e7d45dc5"} pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" containerMessage="Container controller-manager failed liveness probe, will be restarted" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.217691 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" podUID="ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc" containerName="controller-manager" containerID="cri-o://8ef0f4d72f39e4c91f5d814ad0f5248f69fc7641a71cf06a32de5125e7d45dc5" gracePeriod=30 Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.415481 4824 patch_prober.go:28] interesting pod/console-operator-58897d9998-wgtx2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.415540 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" podUID="cd6c14a0-99d8-485d-81f3-cc4ebe5a943b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.427086 4824 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": read tcp 192.168.126.11:39024->192.168.126.11:10257: read: connection reset by peer" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.427162 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": read tcp 192.168.126.11:39024->192.168.126.11:10257: read: connection reset by peer" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.427233 4824 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": unexpected EOF" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.427276 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": unexpected EOF" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.446732 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" event={"ID":"9f87669d-33a9-4269-808c-ca6c718f762c","Type":"ContainerStarted","Data":"872406c61fcebdb3f39a317ce9e250d0ca53bd410abf6cc831134673b957e13b"} Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.450295 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.452561 4824 generic.go:334] "Generic (PLEG): container finished" podID="b0c20e13-b233-4b21-8f47-facef9603735" containerID="715a775b8095fce3f440a6b8d96da8518c365f38a1bcb1424861cf469df16f8a" exitCode=1 Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.452614 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-lfprd" event={"ID":"b0c20e13-b233-4b21-8f47-facef9603735","Type":"ContainerDied","Data":"715a775b8095fce3f440a6b8d96da8518c365f38a1bcb1424861cf469df16f8a"} Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.455476 4824 scope.go:117] "RemoveContainer" containerID="715a775b8095fce3f440a6b8d96da8518c365f38a1bcb1424861cf469df16f8a" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.456926 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-58897d9998-wgtx2_cd6c14a0-99d8-485d-81f3-cc4ebe5a943b/console-operator/0.log" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.456971 4824 generic.go:334] "Generic (PLEG): container finished" podID="cd6c14a0-99d8-485d-81f3-cc4ebe5a943b" containerID="97c7e3b1f76efb699d4015e9514f819e32272fc0f7184743e1eb1a7cea87380e" exitCode=1 Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.457032 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" event={"ID":"cd6c14a0-99d8-485d-81f3-cc4ebe5a943b","Type":"ContainerDied","Data":"97c7e3b1f76efb699d4015e9514f819e32272fc0f7184743e1eb1a7cea87380e"} Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.466008 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" event={"ID":"62823972-71dc-4f7d-b4f6-da0cb90dcf36","Type":"ContainerStarted","Data":"f1cc1168b63ee1dbd664e7d5defcd88202b21989e7ad829670785839ff0a537a"} Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.467090 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.467145 4824 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-9c2mr container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.467174 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" podUID="62823972-71dc-4f7d-b4f6-da0cb90dcf36" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.471365 4824 generic.go:334] "Generic (PLEG): container finished" podID="45732d49-ce14-439b-8b7a-fb9fdf267fb1" containerID="8424dba8744dda5f7cdf5e3a63688e51799cdbdb07d370b2b3b2811ce0148354" exitCode=0 Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.471457 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" event={"ID":"45732d49-ce14-439b-8b7a-fb9fdf267fb1","Type":"ContainerDied","Data":"8424dba8744dda5f7cdf5e3a63688e51799cdbdb07d370b2b3b2811ce0148354"} Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.473521 4824 generic.go:334] "Generic (PLEG): container finished" podID="86c42164-acb8-4a27-9a0f-48307cd304e4" containerID="00bae8ec6f6e4cf5dd4670db60b6420c115bdac303b2c253cc78e718e60f743c" exitCode=0 Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.474404 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" event={"ID":"86c42164-acb8-4a27-9a0f-48307cd304e4","Type":"ContainerDied","Data":"00bae8ec6f6e4cf5dd4670db60b6420c115bdac303b2c253cc78e718e60f743c"} Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.474834 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="webhook-server" containerStatusID={"Type":"cri-o","ID":"499a81ff89ffe2fe094b2e4d1f08ec2b8f670eadfaddc8a27c804fb362e672a6"} pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" containerMessage="Container webhook-server failed liveness probe, will be restarted" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.474883 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" podUID="bab491e7-4eec-4ba1-975d-fb2468372ade" containerName="webhook-server" containerID="cri-o://499a81ff89ffe2fe094b2e4d1f08ec2b8f670eadfaddc8a27c804fb362e672a6" gracePeriod=2 Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.605088 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.705539 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-ntx24" podUID="ad8f524a-8c95-47ab-b74d-9f83331fce76" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.705924 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="metallb-system/frr-k8s-ntx24" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.706032 4824 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-j4zxt container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.706069 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" podUID="7b26f121-04f6-4501-84b2-1833b927aa14" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.707506 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="frr" containerStatusID={"Type":"cri-o","ID":"7123ec58c0ff998d60e21c7177b2ee09325340ea626ba5ccbef31eecc03dd982"} pod="metallb-system/frr-k8s-ntx24" containerMessage="Container frr failed liveness probe, will be restarted" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.707625 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-ntx24" podUID="ad8f524a-8c95-47ab-b74d-9f83331fce76" containerName="frr" containerID="cri-o://7123ec58c0ff998d60e21c7177b2ee09325340ea626ba5ccbef31eecc03dd982" gracePeriod=2 Dec 09 11:14:21 crc kubenswrapper[4824]: E1209 11:14:21.722407 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-56e7587d374a363cda0fabf74a8b632232d4ba0c1278b72f249fc15e66a5e8eb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod612ebf47_a6c4_4a62_8d97_91a003f49c44.slice/crio-6211e2b6e456aecff8e6484685cb1dbfb8ed11c2a63a1ec44ca688b1a82a5c22.scope\": RecentStats: unable to find data in memory cache]" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.743898 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="73d6bd70-44c7-4eed-a93a-36df636869cf" containerName="galera" probeResult="failure" output="command timed out" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.745951 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-k8s-0" podUID="844884cd-6813-4fa2-88e4-7a5994bbd3cb" containerName="prometheus" probeResult="failure" output="command timed out" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.745971 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="844884cd-6813-4fa2-88e4-7a5994bbd3cb" containerName="prometheus" probeResult="failure" output="command timed out" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.832028 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.958994 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" podUID="bab491e7-4eec-4ba1-975d-fb2468372ade" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.95:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.959041 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/controller-f8648f98b-wxgt4" podUID="3af18cd9-4124-4f24-83ea-0ad26ad9ae43" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.97:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.959201 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/controller-f8648f98b-wxgt4" podUID="3af18cd9-4124-4f24-83ea-0ad26ad9ae43" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.97:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.959426 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="metallb-system/controller-f8648f98b-wxgt4" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.959605 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-wxgt4" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.959654 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" podUID="d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/readyz\": dial tcp 10.217.0.114:8081: connect: connection refused" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.960368 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.961019 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="controller" containerStatusID={"Type":"cri-o","ID":"7db44ab4d9712e36348240dbde88e7ceb831da0d9144ade04e056743e6fe40a1"} pod="metallb-system/controller-f8648f98b-wxgt4" containerMessage="Container controller failed liveness probe, will be restarted" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.961119 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/controller-f8648f98b-wxgt4" podUID="3af18cd9-4124-4f24-83ea-0ad26ad9ae43" containerName="controller" containerID="cri-o://7db44ab4d9712e36348240dbde88e7ceb831da0d9144ade04e056743e6fe40a1" gracePeriod=2 Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.969306 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-z68l9" Dec 09 11:14:21 crc kubenswrapper[4824]: I1209 11:14:21.969699 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.034053 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-bf7pj" podUID="ccc192e8-2648-4b1b-8420-1cbd1a27a916" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.034161 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-bf7pj" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.034202 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-wxgt4" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.065841 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.066095 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="54621281-db00-41e4-b617-032435893391" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.0.254:8081/readyz\": dial tcp 10.217.0.254:8081: connect: connection refused" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.090527 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.103324 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.110701 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/prometheus-metric-storage-0" podUID="fb8f1e44-4c21-423a-bacc-1cca7d7715c3" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.167:9090/-/healthy\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.111148 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="fb8f1e44-4c21-423a-bacc-1cca7d7715c3" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.167:9090/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.111391 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.274038 4824 patch_prober.go:28] interesting pod/nmstate-webhook-5f6d4c5ccb-jklpr container/nmstate-webhook namespace/openshift-nmstate: Readiness probe status=failure output="Get \"https://10.217.0.88:9443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.274111 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jklpr" podUID="0b932274-5eba-4e0d-8a64-1f469a6ab3d1" containerName="nmstate-webhook" probeResult="failure" output="Get \"https://10.217.0.88:9443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.285416 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-5xfv2" podUID="87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": dial tcp [::1]:29150: connect: connection refused" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.443009 4824 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.443313 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.486862 4824 generic.go:334] "Generic (PLEG): container finished" podID="143e48cd-c956-448c-8dcb-5858a582e29c" containerID="7681e32f76e067007f5b25ca0d74b0e725c9aaaee11a16e7e0b9b846235bb2a5" exitCode=1 Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.486942 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4" event={"ID":"143e48cd-c956-448c-8dcb-5858a582e29c","Type":"ContainerDied","Data":"7681e32f76e067007f5b25ca0d74b0e725c9aaaee11a16e7e0b9b846235bb2a5"} Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.495758 4824 scope.go:117] "RemoveContainer" containerID="7681e32f76e067007f5b25ca0d74b0e725c9aaaee11a16e7e0b9b846235bb2a5" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.508528 4824 generic.go:334] "Generic (PLEG): container finished" podID="ad8f524a-8c95-47ab-b74d-9f83331fce76" containerID="7123ec58c0ff998d60e21c7177b2ee09325340ea626ba5ccbef31eecc03dd982" exitCode=143 Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.508639 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ntx24" event={"ID":"ad8f524a-8c95-47ab-b74d-9f83331fce76","Type":"ContainerDied","Data":"7123ec58c0ff998d60e21c7177b2ee09325340ea626ba5ccbef31eecc03dd982"} Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.513669 4824 generic.go:334] "Generic (PLEG): container finished" podID="612ebf47-a6c4-4a62-8d97-91a003f49c44" containerID="6211e2b6e456aecff8e6484685cb1dbfb8ed11c2a63a1ec44ca688b1a82a5c22" exitCode=0 Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.515492 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" event={"ID":"612ebf47-a6c4-4a62-8d97-91a003f49c44","Type":"ContainerDied","Data":"6211e2b6e456aecff8e6484685cb1dbfb8ed11c2a63a1ec44ca688b1a82a5c22"} Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.516347 4824 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-9c2mr container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.516398 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" podUID="62823972-71dc-4f7d-b4f6-da0cb90dcf36" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.557058 4824 patch_prober.go:28] interesting pod/metrics-server-65846dd9c4-g8tlp container/metrics-server namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.79:10250/livez\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.557132 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" podUID="e153e30f-4f5e-4a38-8efb-6452096e25d7" containerName="metrics-server" probeResult="failure" output="Get \"https://10.217.0.79:10250/livez\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.561585 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-distributor-76cc67bf56-nl2vt" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.618428 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.622073 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh" podUID="91bf6cd9-a4fe-46e8-b1e7-39e57c1398b3" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/readyz\": read tcp 10.217.0.2:37986->10.217.0.115:8081: read: connection reset by peer" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.758081 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-engine-67ffc9d677-6k2l9" podUID="ce017433-0624-4b3f-b1e9-0de3b591a602" containerName="heat-engine" probeResult="failure" output="command timed out" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.759147 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/heat-engine-67ffc9d677-6k2l9" podUID="ce017433-0624-4b3f-b1e9-0de3b591a602" containerName="heat-engine" probeResult="failure" output="command timed out" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.794033 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-querier-5895d59bb8-89dgf" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.980968 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="153fe73c-2767-4f80-a7b4-0af49257065a" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.208:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.981281 4824 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-k8bgl container/package-server-manager namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.981338 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" podUID="6c10edf9-cf75-4633-8d4c-9bbb9d9b6339" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.981396 4824 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-9c2mr container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.981424 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" podUID="62823972-71dc-4f7d-b4f6-da0cb90dcf36" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.981407 4824 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-tg5lv container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" start-of-body= Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.981455 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" podUID="45732d49-ce14-439b-8b7a-fb9fdf267fb1" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.981599 4824 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-9c2mr container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Dec 09 11:14:22 crc kubenswrapper[4824]: I1209 11:14:22.981619 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" podUID="62823972-71dc-4f7d-b4f6-da0cb90dcf36" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.021888 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/monitoring-plugin-6f74c7bb76-pbw94" Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.076088 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-bf7pj" podUID="ccc192e8-2648-4b1b-8420-1cbd1a27a916" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:14:23 crc kubenswrapper[4824]: E1209 11:14:23.212581 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b70d423ba00ad8d15b6e60bc0af43894068b4f8a5957df3d2fa34df475dfb2cf" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 11:14:23 crc kubenswrapper[4824]: E1209 11:14:23.216413 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b70d423ba00ad8d15b6e60bc0af43894068b4f8a5957df3d2fa34df475dfb2cf" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 11:14:23 crc kubenswrapper[4824]: E1209 11:14:23.220128 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b70d423ba00ad8d15b6e60bc0af43894068b4f8a5957df3d2fa34df475dfb2cf" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 11:14:23 crc kubenswrapper[4824]: E1209 11:14:23.220202 4824 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-marketplace/redhat-operators-582xj" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" containerName="registry-server" Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.305155 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.493199 4824 patch_prober.go:28] interesting pod/router-default-5444994796-mj7jx container/router namespace/openshift-ingress: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]backend-http ok Dec 09 11:14:23 crc kubenswrapper[4824]: [+]has-synced ok Dec 09 11:14:23 crc kubenswrapper[4824]: [-]process-running failed: reason withheld Dec 09 11:14:23 crc kubenswrapper[4824]: healthz check failed Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.493578 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.501966 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-jm5ld" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" containerName="registry-server" probeResult="failure" output="" Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.503632 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/certified-operators-2cbmd" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" containerName="registry-server" probeResult="failure" output="" Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.526887 4824 generic.go:334] "Generic (PLEG): container finished" podID="d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e" containerID="2c66a87b3d637d530e0d9f96512da5a19cae973cc6c8de51cc3f2c64838db824" exitCode=1 Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.527102 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" event={"ID":"d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e","Type":"ContainerDied","Data":"2c66a87b3d637d530e0d9f96512da5a19cae973cc6c8de51cc3f2c64838db824"} Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.528170 4824 scope.go:117] "RemoveContainer" containerID="2c66a87b3d637d530e0d9f96512da5a19cae973cc6c8de51cc3f2c64838db824" Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.549852 4824 generic.go:334] "Generic (PLEG): container finished" podID="23621a98-1d6b-421a-976c-965cecfe6db1" containerID="19c2f889a724f4dacee8917c6176ccbbc55dbfa1bb4f60e0ce6840d832a83568" exitCode=1 Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.550258 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8" event={"ID":"23621a98-1d6b-421a-976c-965cecfe6db1","Type":"ContainerDied","Data":"19c2f889a724f4dacee8917c6176ccbbc55dbfa1bb4f60e0ce6840d832a83568"} Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.551639 4824 scope.go:117] "RemoveContainer" containerID="19c2f889a724f4dacee8917c6176ccbbc55dbfa1bb4f60e0ce6840d832a83568" Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.553551 4824 generic.go:334] "Generic (PLEG): container finished" podID="bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8" containerID="4aecf4910f255e94b17caaa7fae8c9aaa0905e7e1ce503d990addff287a7a433" exitCode=1 Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.553623 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" event={"ID":"bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8","Type":"ContainerDied","Data":"4aecf4910f255e94b17caaa7fae8c9aaa0905e7e1ce503d990addff287a7a433"} Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.554009 4824 scope.go:117] "RemoveContainer" containerID="4aecf4910f255e94b17caaa7fae8c9aaa0905e7e1ce503d990addff287a7a433" Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.560351 4824 generic.go:334] "Generic (PLEG): container finished" podID="383a67da-0f83-41e2-82d6-809d1b6cdbec" containerID="9ea2c6d7fa34ab9c9c15a6234179926d1216a3276cfdc22d1a4626a6ec0c4f5d" exitCode=1 Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.560478 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" event={"ID":"383a67da-0f83-41e2-82d6-809d1b6cdbec","Type":"ContainerDied","Data":"9ea2c6d7fa34ab9c9c15a6234179926d1216a3276cfdc22d1a4626a6ec0c4f5d"} Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.560936 4824 scope.go:117] "RemoveContainer" containerID="9ea2c6d7fa34ab9c9c15a6234179926d1216a3276cfdc22d1a4626a6ec0c4f5d" Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.564295 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/2.log" Dec 09 11:14:23 crc kubenswrapper[4824]: E1209 11:14:23.565855 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1cf855518b8661ca5358ae1fd9e5a18c8e9fddda7448fb2fd36686a961879490 is running failed: container process not found" containerID="1cf855518b8661ca5358ae1fd9e5a18c8e9fddda7448fb2fd36686a961879490" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 11:14:23 crc kubenswrapper[4824]: E1209 11:14:23.566139 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1cf855518b8661ca5358ae1fd9e5a18c8e9fddda7448fb2fd36686a961879490 is running failed: container process not found" containerID="1cf855518b8661ca5358ae1fd9e5a18c8e9fddda7448fb2fd36686a961879490" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 11:14:23 crc kubenswrapper[4824]: E1209 11:14:23.566582 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1cf855518b8661ca5358ae1fd9e5a18c8e9fddda7448fb2fd36686a961879490 is running failed: container process not found" containerID="1cf855518b8661ca5358ae1fd9e5a18c8e9fddda7448fb2fd36686a961879490" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 11:14:23 crc kubenswrapper[4824]: E1209 11:14:23.567130 4824 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1cf855518b8661ca5358ae1fd9e5a18c8e9fddda7448fb2fd36686a961879490 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-95jfr" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" containerName="registry-server" Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.572957 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.574820 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-9d5fm" Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.575068 4824 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="56e7587d374a363cda0fabf74a8b632232d4ba0c1278b72f249fc15e66a5e8eb" exitCode=1 Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.575145 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"56e7587d374a363cda0fabf74a8b632232d4ba0c1278b72f249fc15e66a5e8eb"} Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.575192 4824 scope.go:117] "RemoveContainer" containerID="8b2a31f728d08d316ea20f26b82c72e791d8148fb0e65b5ccd6c881dfaf923e1" Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.576149 4824 scope.go:117] "RemoveContainer" containerID="56e7587d374a363cda0fabf74a8b632232d4ba0c1278b72f249fc15e66a5e8eb" Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.582603 4824 generic.go:334] "Generic (PLEG): container finished" podID="7de1c7e2-adcb-4311-926c-ffc42e3e5fde" containerID="bf6eac2af9b34ed5e8143ec9a4b5e1f1d2a1cf29d103f12c1fa230c24b2a9b8a" exitCode=1 Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.582673 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4q2h7" event={"ID":"7de1c7e2-adcb-4311-926c-ffc42e3e5fde","Type":"ContainerDied","Data":"bf6eac2af9b34ed5e8143ec9a4b5e1f1d2a1cf29d103f12c1fa230c24b2a9b8a"} Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.583831 4824 scope.go:117] "RemoveContainer" containerID="bf6eac2af9b34ed5e8143ec9a4b5e1f1d2a1cf29d103f12c1fa230c24b2a9b8a" Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.589113 4824 generic.go:334] "Generic (PLEG): container finished" podID="6c10edf9-cf75-4633-8d4c-9bbb9d9b6339" containerID="886bc9d74183124b07bb450444597fe02f70436ac80f07efff6ededc8ff84914" exitCode=0 Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.589179 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" event={"ID":"6c10edf9-cf75-4633-8d4c-9bbb9d9b6339","Type":"ContainerDied","Data":"886bc9d74183124b07bb450444597fe02f70436ac80f07efff6ededc8ff84914"} Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.601816 4824 generic.go:334] "Generic (PLEG): container finished" podID="352c7865-5a0b-4ebb-93f2-513ce433ff8f" containerID="e427734a99512a3fca2d9ff58d913920a92fe006a0a467df372b3bea14f49369" exitCode=1 Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.601895 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" event={"ID":"352c7865-5a0b-4ebb-93f2-513ce433ff8f","Type":"ContainerDied","Data":"e427734a99512a3fca2d9ff58d913920a92fe006a0a467df372b3bea14f49369"} Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.603210 4824 scope.go:117] "RemoveContainer" containerID="e427734a99512a3fca2d9ff58d913920a92fe006a0a467df372b3bea14f49369" Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.608437 4824 generic.go:334] "Generic (PLEG): container finished" podID="91bf6cd9-a4fe-46e8-b1e7-39e57c1398b3" containerID="88f94af10301be21fa06f2882918f32b08699007d692a7eaf3f94bdadb4378fe" exitCode=1 Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.608510 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh" event={"ID":"91bf6cd9-a4fe-46e8-b1e7-39e57c1398b3","Type":"ContainerDied","Data":"88f94af10301be21fa06f2882918f32b08699007d692a7eaf3f94bdadb4378fe"} Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.609493 4824 scope.go:117] "RemoveContainer" containerID="88f94af10301be21fa06f2882918f32b08699007d692a7eaf3f94bdadb4378fe" Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.613878 4824 generic.go:334] "Generic (PLEG): container finished" podID="b6577050-6a21-43bb-84aa-20aad247aafc" containerID="1e3e8cc3634bbda54fc7400d8c9ba593261a4d358dd2d6f34df27727570ed602" exitCode=1 Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.613920 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl" event={"ID":"b6577050-6a21-43bb-84aa-20aad247aafc","Type":"ContainerDied","Data":"1e3e8cc3634bbda54fc7400d8c9ba593261a4d358dd2d6f34df27727570ed602"} Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.614779 4824 scope.go:117] "RemoveContainer" containerID="1e3e8cc3634bbda54fc7400d8c9ba593261a4d358dd2d6f34df27727570ed602" Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.807721 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" Dec 09 11:14:23 crc kubenswrapper[4824]: I1209 11:14:23.808093 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.323304 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-compactor-0" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.337812 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.378941 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-ingester-0" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.398137 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.466291 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.641867 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" event={"ID":"d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e","Type":"ContainerStarted","Data":"6d235142cf61f4fc248a5901ae8d09a8571fba41a386d6f8bfafc781fdcaf9b9"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.642164 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.644962 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc" containerID="8ef0f4d72f39e4c91f5d814ad0f5248f69fc7641a71cf06a32de5125e7d45dc5" exitCode=0 Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.645030 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" event={"ID":"ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc","Type":"ContainerDied","Data":"8ef0f4d72f39e4c91f5d814ad0f5248f69fc7641a71cf06a32de5125e7d45dc5"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.649346 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" event={"ID":"45732d49-ce14-439b-8b7a-fb9fdf267fb1","Type":"ContainerStarted","Data":"30d7d1ceff4db8f80cc8c57cefd09488908979adc2f78fd59a77d718975d1cbd"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.649473 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.649521 4824 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-tg5lv container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" start-of-body= Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.649566 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" podUID="45732d49-ce14-439b-8b7a-fb9fdf267fb1" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.660405 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5xfv2" event={"ID":"87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe","Type":"ContainerStarted","Data":"41b1deaf4ca7620c25d7b85c31c340d4370868ccaec854564f8953f00897f632"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.660470 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-5xfv2" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.665650 4824 generic.go:334] "Generic (PLEG): container finished" podID="0af0249c-ed71-4f92-b1a3-cb6f43b92529" containerID="582beb4c6ba000dff1857ea81a561fc444ed9f1574280a492dd1fe4a8ece0cdb" exitCode=1 Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.665727 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-q7psb" event={"ID":"0af0249c-ed71-4f92-b1a3-cb6f43b92529","Type":"ContainerDied","Data":"582beb4c6ba000dff1857ea81a561fc444ed9f1574280a492dd1fe4a8ece0cdb"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.666720 4824 scope.go:117] "RemoveContainer" containerID="582beb4c6ba000dff1857ea81a561fc444ed9f1574280a492dd1fe4a8ece0cdb" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.668901 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-lfprd" event={"ID":"b0c20e13-b233-4b21-8f47-facef9603735","Type":"ContainerStarted","Data":"d128640d258a38c5d3bed5343000dda6eab9191645a64ecbb7be20f5cc0f1e34"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.669440 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-967d97867-lfprd" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.679299 4824 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-j4zxt container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.679358 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" podUID="7b26f121-04f6-4501-84b2-1833b927aa14" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.683523 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4" event={"ID":"143e48cd-c956-448c-8dcb-5858a582e29c","Type":"ContainerStarted","Data":"4d2d22f838f11affb4da66f8bb2255b14dbce10c1962a4d875d69c29ff2073cf"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.684854 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.698338 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" event={"ID":"86c42164-acb8-4a27-9a0f-48307cd304e4","Type":"ContainerStarted","Data":"09d01660c0620dcea94a5723eed579e8ed782ee1f8655c5ca0b8ec041d7820ae"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.699732 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.699825 4824 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-4dlng container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" start-of-body= Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.699852 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" podUID="86c42164-acb8-4a27-9a0f-48307cd304e4" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.733956 4824 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="e1e6cfeb997cafb7ec7f6d829e1ff2e2d95e4b343c422194a5290d222ea53a7a" exitCode=0 Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.734091 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"e1e6cfeb997cafb7ec7f6d829e1ff2e2d95e4b343c422194a5290d222ea53a7a"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.753712 4824 generic.go:334] "Generic (PLEG): container finished" podID="bab491e7-4eec-4ba1-975d-fb2468372ade" containerID="499a81ff89ffe2fe094b2e4d1f08ec2b8f670eadfaddc8a27c804fb362e672a6" exitCode=0 Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.754012 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" event={"ID":"bab491e7-4eec-4ba1-975d-fb2468372ade","Type":"ContainerDied","Data":"499a81ff89ffe2fe094b2e4d1f08ec2b8f670eadfaddc8a27c804fb362e672a6"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.785280 4824 generic.go:334] "Generic (PLEG): container finished" podID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" containerID="4837611c2a862dc0e89e9c677ca72373bccca69d09dbaaa35de3699394b0a701" exitCode=0 Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.785390 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jm5ld" event={"ID":"bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271","Type":"ContainerDied","Data":"4837611c2a862dc0e89e9c677ca72373bccca69d09dbaaa35de3699394b0a701"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.789518 4824 generic.go:334] "Generic (PLEG): container finished" podID="feab57e6-788b-4f23-9e99-aa248786052d" containerID="fe509de64aa6046a4afdd744d7b9a144392cce3237541624889199bb17b79b7b" exitCode=1 Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.789579 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" event={"ID":"feab57e6-788b-4f23-9e99-aa248786052d","Type":"ContainerDied","Data":"fe509de64aa6046a4afdd744d7b9a144392cce3237541624889199bb17b79b7b"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.790516 4824 scope.go:117] "RemoveContainer" containerID="fe509de64aa6046a4afdd744d7b9a144392cce3237541624889199bb17b79b7b" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.792478 4824 generic.go:334] "Generic (PLEG): container finished" podID="3af18cd9-4124-4f24-83ea-0ad26ad9ae43" containerID="7db44ab4d9712e36348240dbde88e7ceb831da0d9144ade04e056743e6fe40a1" exitCode=0 Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.792540 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-wxgt4" event={"ID":"3af18cd9-4124-4f24-83ea-0ad26ad9ae43","Type":"ContainerDied","Data":"7db44ab4d9712e36348240dbde88e7ceb831da0d9144ade04e056743e6fe40a1"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.794533 4824 generic.go:334] "Generic (PLEG): container finished" podID="7b26f121-04f6-4501-84b2-1833b927aa14" containerID="3efabae52732fd1dd695c8ec9ef93d6237ea498a410bac3039995980f09cd0d7" exitCode=0 Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.794582 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" event={"ID":"7b26f121-04f6-4501-84b2-1833b927aa14","Type":"ContainerDied","Data":"3efabae52732fd1dd695c8ec9ef93d6237ea498a410bac3039995980f09cd0d7"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.795887 4824 generic.go:334] "Generic (PLEG): container finished" podID="c3400b7c-5666-4b36-a50f-7800f6737527" containerID="ededff8a2f4ee4e72db8028a40598edd0ce0981b53e5b3d8e13b10053bf64cd1" exitCode=0 Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.795928 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" event={"ID":"c3400b7c-5666-4b36-a50f-7800f6737527","Type":"ContainerDied","Data":"ededff8a2f4ee4e72db8028a40598edd0ce0981b53e5b3d8e13b10053bf64cd1"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.821903 4824 generic.go:334] "Generic (PLEG): container finished" podID="f2572cf3-89f1-49d7-9246-9350703e9764" containerID="b70d423ba00ad8d15b6e60bc0af43894068b4f8a5957df3d2fa34df475dfb2cf" exitCode=0 Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.821974 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-582xj" event={"ID":"f2572cf3-89f1-49d7-9246-9350703e9764","Type":"ContainerDied","Data":"b70d423ba00ad8d15b6e60bc0af43894068b4f8a5957df3d2fa34df475dfb2cf"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.835296 4824 generic.go:334] "Generic (PLEG): container finished" podID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" containerID="8eb26eb28020a6b331e841e6bc1a33e26e08169614eddc4b7de37f69670eabf9" exitCode=0 Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.835394 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cbmd" event={"ID":"257d6d77-75f1-41e0-8bd2-c6ec3c1ff582","Type":"ContainerDied","Data":"8eb26eb28020a6b331e841e6bc1a33e26e08169614eddc4b7de37f69670eabf9"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.839860 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-58897d9998-wgtx2_cd6c14a0-99d8-485d-81f3-cc4ebe5a943b/console-operator/0.log" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.839967 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" event={"ID":"cd6c14a0-99d8-485d-81f3-cc4ebe5a943b","Type":"ContainerStarted","Data":"959da6351c8b1617a623578112a1dc7825d9f71cc3dd6e417553f2a0c5b2f0be"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.840466 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.840541 4824 patch_prober.go:28] interesting pod/console-operator-58897d9998-wgtx2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.840569 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" podUID="cd6c14a0-99d8-485d-81f3-cc4ebe5a943b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.843546 4824 generic.go:334] "Generic (PLEG): container finished" podID="afe92d2b-bce3-445e-9c26-9c533353e7a3" containerID="03c338c0a6a98a41c144c02fffabcfe6e5a3daf642e4e2523e58122af8753be4" exitCode=0 Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.843557 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" event={"ID":"afe92d2b-bce3-445e-9c26-9c533353e7a3","Type":"ContainerDied","Data":"03c338c0a6a98a41c144c02fffabcfe6e5a3daf642e4e2523e58122af8753be4"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.853341 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c2dd0680-d32f-45e8-b834-6a327d2eaa21/ovn-northd/0.log" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.853441 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c2dd0680-d32f-45e8-b834-6a327d2eaa21","Type":"ContainerStarted","Data":"6c7561fbd27950b0f8c14d370b98e686bf4eb75fac6912631d5fbe5b1096149a"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.855106 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.878363 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-598df84fd8-l2bt7" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.878606 4824 generic.go:334] "Generic (PLEG): container finished" podID="c2889de3-49b2-4465-8235-37f9e58b42a3" containerID="c4a6bd38a195f4eb3c912c97bc78383959a4e5d35e8611606e390dd61f32d99e" exitCode=1 Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.878651 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" event={"ID":"c2889de3-49b2-4465-8235-37f9e58b42a3","Type":"ContainerDied","Data":"c4a6bd38a195f4eb3c912c97bc78383959a4e5d35e8611606e390dd61f32d99e"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.879766 4824 scope.go:117] "RemoveContainer" containerID="c4a6bd38a195f4eb3c912c97bc78383959a4e5d35e8611606e390dd61f32d99e" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.907793 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_router-default-5444994796-mj7jx_dac1ac9b-4ddd-41b2-813b-ed7d8a71642f/router/0.log" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.907865 4824 generic.go:334] "Generic (PLEG): container finished" podID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerID="b81f987a695aac3b20392984022ac9b8c31c087c6d4439490168cb2c88e811c7" exitCode=137 Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.907989 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-mj7jx" event={"ID":"dac1ac9b-4ddd-41b2-813b-ed7d8a71642f","Type":"ContainerDied","Data":"b81f987a695aac3b20392984022ac9b8c31c087c6d4439490168cb2c88e811c7"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.936562 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.938483 4824 generic.go:334] "Generic (PLEG): container finished" podID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" containerID="1cf855518b8661ca5358ae1fd9e5a18c8e9fddda7448fb2fd36686a961879490" exitCode=0 Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.938542 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-95jfr" event={"ID":"6ffd99b7-9bb8-479e-a1a1-7f31ee006e38","Type":"ContainerDied","Data":"1cf855518b8661ca5358ae1fd9e5a18c8e9fddda7448fb2fd36686a961879490"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.982236 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ntx24" event={"ID":"ad8f524a-8c95-47ab-b74d-9f83331fce76","Type":"ContainerStarted","Data":"b62e910d7ef7e21f3907e48d72d25d810d091642913b5686816a33a55973236d"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.988027 4824 generic.go:334] "Generic (PLEG): container finished" podID="6e91c44f-6221-4dfd-9c18-2bbedff02850" containerID="f0b3f37bdad3f939e038e3d3ace05555dbd1fba79e301122328cdedead3df512" exitCode=1 Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.988082 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl" event={"ID":"6e91c44f-6221-4dfd-9c18-2bbedff02850","Type":"ContainerDied","Data":"f0b3f37bdad3f939e038e3d3ace05555dbd1fba79e301122328cdedead3df512"} Dec 09 11:14:24 crc kubenswrapper[4824]: I1209 11:14:24.989339 4824 scope.go:117] "RemoveContainer" containerID="f0b3f37bdad3f939e038e3d3ace05555dbd1fba79e301122328cdedead3df512" Dec 09 11:14:25 crc kubenswrapper[4824]: E1209 11:14:25.311926 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4837611c2a862dc0e89e9c677ca72373bccca69d09dbaaa35de3699394b0a701 is running failed: container process not found" containerID="4837611c2a862dc0e89e9c677ca72373bccca69d09dbaaa35de3699394b0a701" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 11:14:25 crc kubenswrapper[4824]: E1209 11:14:25.312687 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4837611c2a862dc0e89e9c677ca72373bccca69d09dbaaa35de3699394b0a701 is running failed: container process not found" containerID="4837611c2a862dc0e89e9c677ca72373bccca69d09dbaaa35de3699394b0a701" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 11:14:25 crc kubenswrapper[4824]: E1209 11:14:25.313687 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4837611c2a862dc0e89e9c677ca72373bccca69d09dbaaa35de3699394b0a701 is running failed: container process not found" containerID="4837611c2a862dc0e89e9c677ca72373bccca69d09dbaaa35de3699394b0a701" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 11:14:25 crc kubenswrapper[4824]: E1209 11:14:25.313719 4824 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4837611c2a862dc0e89e9c677ca72373bccca69d09dbaaa35de3699394b0a701 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-jm5ld" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" containerName="registry-server" Dec 09 11:14:25 crc kubenswrapper[4824]: I1209 11:14:25.656083 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="73d6bd70-44c7-4eed-a93a-36df636869cf" containerName="galera" containerID="cri-o://3187d927bee90adbfb095796189ae710aa8a824436cd0f34583d4da240bacb9d" gracePeriod=26 Dec 09 11:14:25 crc kubenswrapper[4824]: I1209 11:14:25.663202 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-ntx24" Dec 09 11:14:25 crc kubenswrapper[4824]: I1209 11:14:25.715537 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-ntx24" Dec 09 11:14:25 crc kubenswrapper[4824]: I1209 11:14:25.790307 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="56148e0f-636f-410e-bfb5-342da01e8c76" containerName="galera" containerID="cri-o://4f85804d9b180ab6d108fd3769e0a57f89c384076fef52095c519bf02c5c868e" gracePeriod=14 Dec 09 11:14:25 crc kubenswrapper[4824]: E1209 11:14:25.899235 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8eb26eb28020a6b331e841e6bc1a33e26e08169614eddc4b7de37f69670eabf9 is running failed: container process not found" containerID="8eb26eb28020a6b331e841e6bc1a33e26e08169614eddc4b7de37f69670eabf9" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 11:14:25 crc kubenswrapper[4824]: E1209 11:14:25.903665 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8eb26eb28020a6b331e841e6bc1a33e26e08169614eddc4b7de37f69670eabf9 is running failed: container process not found" containerID="8eb26eb28020a6b331e841e6bc1a33e26e08169614eddc4b7de37f69670eabf9" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 11:14:25 crc kubenswrapper[4824]: E1209 11:14:25.904890 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8eb26eb28020a6b331e841e6bc1a33e26e08169614eddc4b7de37f69670eabf9 is running failed: container process not found" containerID="8eb26eb28020a6b331e841e6bc1a33e26e08169614eddc4b7de37f69670eabf9" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 11:14:25 crc kubenswrapper[4824]: E1209 11:14:25.904943 4824 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8eb26eb28020a6b331e841e6bc1a33e26e08169614eddc4b7de37f69670eabf9 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-2cbmd" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" containerName="registry-server" Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.026280 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh" event={"ID":"91bf6cd9-a4fe-46e8-b1e7-39e57c1398b3","Type":"ContainerStarted","Data":"7303ef1edb111fcbe910811ce5fa12ff29e346f96e9d3d8b464ca000af7f099e"} Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.028138 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh" Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.035978 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" event={"ID":"352c7865-5a0b-4ebb-93f2-513ce433ff8f","Type":"ContainerStarted","Data":"8bc19e9a1e4ac657a0eb38e7d4c3df469bfc33710342fbae996c94d3caf8fe41"} Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.036493 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.053744 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8" event={"ID":"23621a98-1d6b-421a-976c-965cecfe6db1","Type":"ContainerStarted","Data":"3c882f54d5e1b95e0aea89a5eab335b640dd7fa7667e4e3438a43866f813a466"} Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.055108 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8" Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.056551 4824 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": dial tcp 192.168.126.11:10259: connect: connection refused" start-of-body= Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.056593 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": dial tcp 192.168.126.11:10259: connect: connection refused" Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.098737 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" event={"ID":"383a67da-0f83-41e2-82d6-809d1b6cdbec","Type":"ContainerStarted","Data":"0fda2077e038c2cbca169040274baf173b3156253b65c4c7535af4c12231ddb5"} Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.104427 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.121249 4824 generic.go:334] "Generic (PLEG): container finished" podID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" exitCode=0 Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.121352 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9339c10c-1cea-4ffc-8931-9b349a1e74cd","Type":"ContainerDied","Data":"c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd"} Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.126050 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl" event={"ID":"b6577050-6a21-43bb-84aa-20aad247aafc","Type":"ContainerStarted","Data":"82bee8164d84cd6a289900152cfeaa2ea62cca2e8d4107d8f6e926d626870d3a"} Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.127465 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl" Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.138666 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1e8ecef7647993de3ec58819f3cf73633407ee22c027ae2261bd146be93f4ad1"} Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.139735 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.142000 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/2.log" Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.145258 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c50ac99b04609c28c81b4f5a019f7955d85242aa17bf45fae531ccd9c7443e2e"} Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.150940 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4q2h7" event={"ID":"7de1c7e2-adcb-4311-926c-ffc42e3e5fde","Type":"ContainerStarted","Data":"442e67b5cc457373dbc43bd19a52a51a79759f88ddd5967a9915231d32e97040"} Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.151436 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4q2h7" Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.159720 4824 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-4dlng container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" start-of-body= Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.159771 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" podUID="86c42164-acb8-4a27-9a0f-48307cd304e4" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.159869 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" event={"ID":"bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8","Type":"ContainerStarted","Data":"26b791d1793ac85c5038106436f1ef9409b52ffeabb679c7838d05f0792080a7"} Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.160384 4824 patch_prober.go:28] interesting pod/console-operator-58897d9998-wgtx2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.160438 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" podUID="cd6c14a0-99d8-485d-81f3-cc4ebe5a943b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.160718 4824 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-tg5lv container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" start-of-body= Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.160811 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" podUID="45732d49-ce14-439b-8b7a-fb9fdf267fb1" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.161478 4824 status_manager.go:317] "Container readiness changed for unknown container" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" containerID="cri-o://4aecf4910f255e94b17caaa7fae8c9aaa0905e7e1ce503d990addff287a7a433" Dec 09 11:14:26 crc kubenswrapper[4824]: I1209 11:14:26.161498 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" Dec 09 11:14:26 crc kubenswrapper[4824]: E1209 11:14:26.381555 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 11:14:26 crc kubenswrapper[4824]: E1209 11:14:26.381612 4824 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 11:14:26 crc kubenswrapper[4824]: E1209 11:14:26.382255 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n9h55h68ch65bh655h64chbfh6ch5b4h575h6bh654h5bdh587h5h58bh5dch5b9hf4h548h579hf8h5f5h686h667h97h57hb5h65h656h588h5f8q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6p7fx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9339c10c-1cea-4ffc-8931-9b349a1e74cd): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 11:14:26 crc kubenswrapper[4824]: E1209 11:14:26.383520 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.023394 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-p5zwg" Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.179630 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" event={"ID":"feab57e6-788b-4f23-9e99-aa248786052d","Type":"ContainerStarted","Data":"a15c165e33fe8d347277d973290ba6eaed98975827a5b2df81550cc7cae70d02"} Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.188892 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" event={"ID":"c2889de3-49b2-4465-8235-37f9e58b42a3","Type":"ContainerStarted","Data":"10e41c0a2945d352811489259960ce312aea2906ae293ac20473ddb3c377d5e3"} Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.206611 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-wxgt4" event={"ID":"3af18cd9-4124-4f24-83ea-0ad26ad9ae43","Type":"ContainerStarted","Data":"56d10235a2c45f7f2f629bde75ce783be4c4722b013324aa3836b1640a17e58f"} Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.208543 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-wxgt4" Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.222583 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-bxplc" event={"ID":"c3400b7c-5666-4b36-a50f-7800f6737527","Type":"ContainerStarted","Data":"bd7865cd94b8891843879d83460bd5e25bfd4bcd6a66d5f45673f429373ba13f"} Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.235214 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" podUID="9f87669d-33a9-4269-808c-ca6c718f762c" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.44:6080/healthz\": dial tcp 10.217.0.44:6080: connect: connection refused" Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.236715 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" event={"ID":"612ebf47-a6c4-4a62-8d97-91a003f49c44","Type":"ContainerStarted","Data":"5d56fb8561aa4fbf470bc64005d0efddd8896139229b82513eb7bcb87c783766"} Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.238261 4824 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-qn2qz container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.5:8081/healthz\": dial tcp 10.217.0.5:8081: connect: connection refused" start-of-body= Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.238302 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" podUID="612ebf47-a6c4-4a62-8d97-91a003f49c44" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.5:8081/healthz\": dial tcp 10.217.0.5:8081: connect: connection refused" Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.238365 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.241674 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" event={"ID":"afe92d2b-bce3-445e-9c26-9c533353e7a3","Type":"ContainerStarted","Data":"7b27a9a5b4a418d61eb7f7df3f5cc036ef08c4959cb60ad3237fcfaf80ffb5a1"} Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.242023 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.243059 4824 patch_prober.go:28] interesting pod/route-controller-manager-7f84c5b8f9-rqmr9 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" start-of-body= Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.243104 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" podUID="afe92d2b-bce3-445e-9c26-9c533353e7a3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.249685 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" event={"ID":"6c10edf9-cf75-4633-8d4c-9bbb9d9b6339","Type":"ContainerStarted","Data":"dee6b8520c585d123c1f4f2860334f45f2e2352544da2c15caa0af3a9fe7411a"} Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.250490 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.253081 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl" event={"ID":"6e91c44f-6221-4dfd-9c18-2bbedff02850","Type":"ContainerStarted","Data":"d0c7b93e5964ff3e9e1fbb187254bf9f2b1fc5bb8756dbc969bb098f7efc00bf"} Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.253411 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl" Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.279218 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" event={"ID":"ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc","Type":"ContainerStarted","Data":"d6aefab02fc86b7869966d89d14c96701ad792db1023a254e19b309072c1cf4e"} Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.279266 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.283927 4824 patch_prober.go:28] interesting pod/controller-manager-69949c7bd4-pmmk2 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.78:8443/healthz\": dial tcp 10.217.0.78:8443: connect: connection refused" start-of-body= Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.283988 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" podUID="ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.78:8443/healthz\": dial tcp 10.217.0.78:8443: connect: connection refused" Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.310599 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" event={"ID":"7b26f121-04f6-4501-84b2-1833b927aa14","Type":"ContainerStarted","Data":"be714e11dae0b87c25d5582984ab647e63588f36b7e2976cec5c6723aa5b5fd7"} Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.312019 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.333318 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_router-default-5444994796-mj7jx_dac1ac9b-4ddd-41b2-813b-ed7d8a71642f/router/0.log" Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.336429 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.337822 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-mj7jx" event={"ID":"dac1ac9b-4ddd-41b2-813b-ed7d8a71642f","Type":"ContainerStarted","Data":"4ca29e1a614c1c80e1394b4b3a4bee2fd37d94f5415658447e385cb8cff9a85c"} Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.342583 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.427236 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-4dlng" Dec 09 11:14:27 crc kubenswrapper[4824]: E1209 11:14:27.454257 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.549560 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.549900 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" Dec 09 11:14:27 crc kubenswrapper[4824]: E1209 11:14:27.749817 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4f85804d9b180ab6d108fd3769e0a57f89c384076fef52095c519bf02c5c868e" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 09 11:14:27 crc kubenswrapper[4824]: E1209 11:14:27.764223 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4f85804d9b180ab6d108fd3769e0a57f89c384076fef52095c519bf02c5c868e" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 09 11:14:27 crc kubenswrapper[4824]: E1209 11:14:27.773225 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4f85804d9b180ab6d108fd3769e0a57f89c384076fef52095c519bf02c5c868e" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 09 11:14:27 crc kubenswrapper[4824]: E1209 11:14:27.773312 4824 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="56148e0f-636f-410e-bfb5-342da01e8c76" containerName="galera" Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.839264 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.840243 4824 patch_prober.go:28] interesting pod/router-default-5444994796-mj7jx container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Dec 09 11:14:27 crc kubenswrapper[4824]: I1209 11:14:27.840285 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Dec 09 11:14:28 crc kubenswrapper[4824]: I1209 11:14:28.305238 4824 patch_prober.go:28] interesting pod/route-controller-manager-7f84c5b8f9-rqmr9 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" start-of-body= Dec 09 11:14:28 crc kubenswrapper[4824]: I1209 11:14:28.305517 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" podUID="afe92d2b-bce3-445e-9c26-9c533353e7a3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" Dec 09 11:14:28 crc kubenswrapper[4824]: I1209 11:14:28.352157 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-q7psb" event={"ID":"0af0249c-ed71-4f92-b1a3-cb6f43b92529","Type":"ContainerStarted","Data":"34a50db41e471ecba8bbe3bc00224d6cdf50077bf8b12a3522137feacb73ca6c"} Dec 09 11:14:28 crc kubenswrapper[4824]: I1209 11:14:28.352932 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-q7psb" Dec 09 11:14:28 crc kubenswrapper[4824]: I1209 11:14:28.359937 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" event={"ID":"bab491e7-4eec-4ba1-975d-fb2468372ade","Type":"ContainerStarted","Data":"f80463cc54fba57fcfa680fbf7b8c8a5ce5b4cb977eddcef766776c91cf8e528"} Dec 09 11:14:28 crc kubenswrapper[4824]: I1209 11:14:28.362769 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"54621281-db00-41e4-b617-032435893391","Type":"ContainerStarted","Data":"ede1311186cfe6256177ca91e976ddbfe440f5bf60b7be27f4f530c15e8f019b"} Dec 09 11:14:28 crc kubenswrapper[4824]: I1209 11:14:28.370479 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jm5ld" event={"ID":"bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271","Type":"ContainerStarted","Data":"422e7110c8297684947e4218beab74c2802cb4e21bbd38d973bcbf390774824c"} Dec 09 11:14:28 crc kubenswrapper[4824]: I1209 11:14:28.371221 4824 patch_prober.go:28] interesting pod/route-controller-manager-7f84c5b8f9-rqmr9 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" start-of-body= Dec 09 11:14:28 crc kubenswrapper[4824]: I1209 11:14:28.371264 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" podUID="afe92d2b-bce3-445e-9c26-9c533353e7a3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" Dec 09 11:14:28 crc kubenswrapper[4824]: I1209 11:14:28.371867 4824 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-qn2qz container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.5:8081/healthz\": dial tcp 10.217.0.5:8081: connect: connection refused" start-of-body= Dec 09 11:14:28 crc kubenswrapper[4824]: I1209 11:14:28.371928 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" podUID="612ebf47-a6c4-4a62-8d97-91a003f49c44" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.5:8081/healthz\": dial tcp 10.217.0.5:8081: connect: connection refused" Dec 09 11:14:28 crc kubenswrapper[4824]: I1209 11:14:28.372648 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" Dec 09 11:14:28 crc kubenswrapper[4824]: I1209 11:14:28.372688 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" Dec 09 11:14:28 crc kubenswrapper[4824]: I1209 11:14:28.373105 4824 patch_prober.go:28] interesting pod/controller-manager-69949c7bd4-pmmk2 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.78:8443/healthz\": dial tcp 10.217.0.78:8443: connect: connection refused" start-of-body= Dec 09 11:14:28 crc kubenswrapper[4824]: I1209 11:14:28.373143 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" podUID="ee9b1e8c-b0c9-47ee-8c98-e8fe5153b6bc" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.78:8443/healthz\": dial tcp 10.217.0.78:8443: connect: connection refused" Dec 09 11:14:28 crc kubenswrapper[4824]: I1209 11:14:28.839391 4824 patch_prober.go:28] interesting pod/router-default-5444994796-mj7jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:14:28 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Dec 09 11:14:28 crc kubenswrapper[4824]: [+]process-running ok Dec 09 11:14:28 crc kubenswrapper[4824]: healthz check failed Dec 09 11:14:28 crc kubenswrapper[4824]: I1209 11:14:28.839663 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:14:29 crc kubenswrapper[4824]: E1209 11:14:29.350967 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3187d927bee90adbfb095796189ae710aa8a824436cd0f34583d4da240bacb9d" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 09 11:14:29 crc kubenswrapper[4824]: E1209 11:14:29.360213 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3187d927bee90adbfb095796189ae710aa8a824436cd0f34583d4da240bacb9d" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 09 11:14:29 crc kubenswrapper[4824]: E1209 11:14:29.361841 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3187d927bee90adbfb095796189ae710aa8a824436cd0f34583d4da240bacb9d" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 09 11:14:29 crc kubenswrapper[4824]: E1209 11:14:29.361929 4824 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="73d6bd70-44c7-4eed-a93a-36df636869cf" containerName="galera" Dec 09 11:14:29 crc kubenswrapper[4824]: I1209 11:14:29.383156 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" Dec 09 11:14:29 crc kubenswrapper[4824]: I1209 11:14:29.384013 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" Dec 09 11:14:29 crc kubenswrapper[4824]: I1209 11:14:29.841315 4824 patch_prober.go:28] interesting pod/router-default-5444994796-mj7jx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 11:14:29 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Dec 09 11:14:29 crc kubenswrapper[4824]: [+]process-running ok Dec 09 11:14:29 crc kubenswrapper[4824]: healthz check failed Dec 09 11:14:29 crc kubenswrapper[4824]: I1209 11:14:29.841418 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mj7jx" podUID="dac1ac9b-4ddd-41b2-813b-ed7d8a71642f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:14:29 crc kubenswrapper[4824]: I1209 11:14:29.859432 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-qn2qz" Dec 09 11:14:30 crc kubenswrapper[4824]: I1209 11:14:30.139419 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-69949c7bd4-pmmk2" Dec 09 11:14:30 crc kubenswrapper[4824]: I1209 11:14:30.250892 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-j4zxt" Dec 09 11:14:30 crc kubenswrapper[4824]: I1209 11:14:30.313097 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="153fe73c-2767-4f80-a7b4-0af49257065a" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:14:30 crc kubenswrapper[4824]: I1209 11:14:30.396281 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cbmd" event={"ID":"257d6d77-75f1-41e0-8bd2-c6ec3c1ff582","Type":"ContainerStarted","Data":"50b5a59aa3cff21c7dd0d8589aef0b16cb173989466b6fbbed3356da31c9ce86"} Dec 09 11:14:30 crc kubenswrapper[4824]: I1209 11:14:30.397211 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="54621281-db00-41e4-b617-032435893391" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.0.254:8081/readyz\": dial tcp 10.217.0.254:8081: connect: connection refused" Dec 09 11:14:30 crc kubenswrapper[4824]: I1209 11:14:30.840009 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 11:14:30 crc kubenswrapper[4824]: I1209 11:14:30.840655 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 11:14:30 crc kubenswrapper[4824]: I1209 11:14:30.856536 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-mj7jx" Dec 09 11:14:30 crc kubenswrapper[4824]: I1209 11:14:30.880341 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-n45n4" Dec 09 11:14:30 crc kubenswrapper[4824]: I1209 11:14:30.941456 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="153fe73c-2767-4f80-a7b4-0af49257065a" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:14:30 crc kubenswrapper[4824]: I1209 11:14:30.992741 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-bf7pj" Dec 09 11:14:31 crc kubenswrapper[4824]: I1209 11:14:31.208145 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-967d97867-lfprd" Dec 09 11:14:31 crc kubenswrapper[4824]: I1209 11:14:31.274843 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4q2h7" Dec 09 11:14:31 crc kubenswrapper[4824]: I1209 11:14:31.410060 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-582xj" event={"ID":"f2572cf3-89f1-49d7-9246-9350703e9764","Type":"ContainerStarted","Data":"b17345a1387fa7bc676fa7b0f8dbb6d84bd5269210d0453e07d3fbb65eb9bc6b"} Dec 09 11:14:31 crc kubenswrapper[4824]: I1209 11:14:31.483468 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-wgtx2" Dec 09 11:14:31 crc kubenswrapper[4824]: I1209 11:14:31.490048 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-r9fcl" Dec 09 11:14:31 crc kubenswrapper[4824]: I1209 11:14:31.834756 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hlwv8" Dec 09 11:14:31 crc kubenswrapper[4824]: I1209 11:14:31.862134 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rbwkh" Dec 09 11:14:31 crc kubenswrapper[4824]: I1209 11:14:31.889664 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5vnrh" Dec 09 11:14:31 crc kubenswrapper[4824]: I1209 11:14:31.899054 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-pqhrq" Dec 09 11:14:31 crc kubenswrapper[4824]: I1209 11:14:31.968684 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-rz4sl" Dec 09 11:14:32 crc kubenswrapper[4824]: I1209 11:14:32.066570 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 09 11:14:32 crc kubenswrapper[4824]: I1209 11:14:32.070302 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-577c7bd649-xtnkl" Dec 09 11:14:32 crc kubenswrapper[4824]: I1209 11:14:32.092770 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-55kwm" Dec 09 11:14:32 crc kubenswrapper[4824]: I1209 11:14:32.108366 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-q7psb" Dec 09 11:14:32 crc kubenswrapper[4824]: E1209 11:14:32.155520 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfeab57e6_788b_4f23_9e99_aa248786052d.slice/crio-fe509de64aa6046a4afdd744d7b9a144392cce3237541624889199bb17b79b7b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee9b1e8c_b0c9_47ee_8c98_e8fe5153b6bc.slice/crio-conmon-8ef0f4d72f39e4c91f5d814ad0f5248f69fc7641a71cf06a32de5125e7d45dc5.scope\": RecentStats: unable to find data in memory cache]" Dec 09 11:14:32 crc kubenswrapper[4824]: I1209 11:14:32.234431 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-dbnmx" Dec 09 11:14:32 crc kubenswrapper[4824]: I1209 11:14:32.299137 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 09 11:14:32 crc kubenswrapper[4824]: I1209 11:14:32.425312 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-95jfr" event={"ID":"6ffd99b7-9bb8-479e-a1a1-7f31ee006e38","Type":"ContainerStarted","Data":"8f8ecda72950ef03ff3fbe1a14b0f63ae3767ae12a16e550b24761d07e9d0b05"} Dec 09 11:14:32 crc kubenswrapper[4824]: I1209 11:14:32.957453 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg5lv" Dec 09 11:14:32 crc kubenswrapper[4824]: I1209 11:14:32.980728 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9c2mr" Dec 09 11:14:33 crc kubenswrapper[4824]: I1209 11:14:33.201818 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-582xj" Dec 09 11:14:33 crc kubenswrapper[4824]: I1209 11:14:33.201868 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-582xj" Dec 09 11:14:33 crc kubenswrapper[4824]: I1209 11:14:33.438820 4824 generic.go:334] "Generic (PLEG): container finished" podID="73d6bd70-44c7-4eed-a93a-36df636869cf" containerID="3187d927bee90adbfb095796189ae710aa8a824436cd0f34583d4da240bacb9d" exitCode=0 Dec 09 11:14:33 crc kubenswrapper[4824]: I1209 11:14:33.439124 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"73d6bd70-44c7-4eed-a93a-36df636869cf","Type":"ContainerDied","Data":"3187d927bee90adbfb095796189ae710aa8a824436cd0f34583d4da240bacb9d"} Dec 09 11:14:33 crc kubenswrapper[4824]: I1209 11:14:33.439156 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"73d6bd70-44c7-4eed-a93a-36df636869cf","Type":"ContainerStarted","Data":"603481d0f15e5e75450c666f68798c3c62eb68b9bf9e98747145e8b16124eb5b"} Dec 09 11:14:33 crc kubenswrapper[4824]: I1209 11:14:33.565881 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-95jfr" Dec 09 11:14:33 crc kubenswrapper[4824]: I1209 11:14:33.565944 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-95jfr" Dec 09 11:14:33 crc kubenswrapper[4824]: I1209 11:14:33.810547 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7" Dec 09 11:14:33 crc kubenswrapper[4824]: I1209 11:14:33.942521 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="153fe73c-2767-4f80-a7b4-0af49257065a" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:14:33 crc kubenswrapper[4824]: I1209 11:14:33.942608 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 09 11:14:33 crc kubenswrapper[4824]: I1209 11:14:33.943708 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cinder-scheduler" containerStatusID={"Type":"cri-o","ID":"d2f14825848e7e355909e904cb07d3ac37df2eec491e11d6f97757df6a0f4a7d"} pod="openstack/cinder-scheduler-0" containerMessage="Container cinder-scheduler failed liveness probe, will be restarted" Dec 09 11:14:33 crc kubenswrapper[4824]: I1209 11:14:33.943774 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="153fe73c-2767-4f80-a7b4-0af49257065a" containerName="cinder-scheduler" containerID="cri-o://d2f14825848e7e355909e904cb07d3ac37df2eec491e11d6f97757df6a0f4a7d" gracePeriod=30 Dec 09 11:14:34 crc kubenswrapper[4824]: I1209 11:14:34.277409 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-582xj" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" containerName="registry-server" probeResult="failure" output=< Dec 09 11:14:34 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 11:14:34 crc kubenswrapper[4824]: > Dec 09 11:14:34 crc kubenswrapper[4824]: I1209 11:14:34.337753 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:14:34 crc kubenswrapper[4824]: I1209 11:14:34.620416 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-95jfr" podUID="6ffd99b7-9bb8-479e-a1a1-7f31ee006e38" containerName="registry-server" probeResult="failure" output=< Dec 09 11:14:34 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 11:14:34 crc kubenswrapper[4824]: > Dec 09 11:14:34 crc kubenswrapper[4824]: I1209 11:14:34.937179 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:14:34 crc kubenswrapper[4824]: I1209 11:14:34.941508 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:14:35 crc kubenswrapper[4824]: I1209 11:14:35.188760 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 09 11:14:35 crc kubenswrapper[4824]: I1209 11:14:35.304407 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jm5ld" Dec 09 11:14:35 crc kubenswrapper[4824]: I1209 11:14:35.304465 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jm5ld" Dec 09 11:14:35 crc kubenswrapper[4824]: I1209 11:14:35.496960 4824 generic.go:334] "Generic (PLEG): container finished" podID="0fd5955c-4039-42a1-8ecd-0256bb683e61" containerID="a8a574116ea94300bb3b084f452cedabd53a1b7b6e69d43cda4dd982221f00fe" exitCode=1 Dec 09 11:14:35 crc kubenswrapper[4824]: I1209 11:14:35.497087 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"0fd5955c-4039-42a1-8ecd-0256bb683e61","Type":"ContainerDied","Data":"a8a574116ea94300bb3b084f452cedabd53a1b7b6e69d43cda4dd982221f00fe"} Dec 09 11:14:35 crc kubenswrapper[4824]: I1209 11:14:35.512531 4824 generic.go:334] "Generic (PLEG): container finished" podID="56148e0f-636f-410e-bfb5-342da01e8c76" containerID="4f85804d9b180ab6d108fd3769e0a57f89c384076fef52095c519bf02c5c868e" exitCode=0 Dec 09 11:14:35 crc kubenswrapper[4824]: I1209 11:14:35.513891 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"56148e0f-636f-410e-bfb5-342da01e8c76","Type":"ContainerDied","Data":"4f85804d9b180ab6d108fd3769e0a57f89c384076fef52095c519bf02c5c868e"} Dec 09 11:14:35 crc kubenswrapper[4824]: I1209 11:14:35.527284 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 11:14:35 crc kubenswrapper[4824]: I1209 11:14:35.892102 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2cbmd" Dec 09 11:14:35 crc kubenswrapper[4824]: I1209 11:14:35.893112 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2cbmd" Dec 09 11:14:36 crc kubenswrapper[4824]: I1209 11:14:36.372811 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-jm5ld" podUID="bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271" containerName="registry-server" probeResult="failure" output=< Dec 09 11:14:36 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 11:14:36 crc kubenswrapper[4824]: > Dec 09 11:14:36 crc kubenswrapper[4824]: I1209 11:14:36.527224 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"56148e0f-636f-410e-bfb5-342da01e8c76","Type":"ContainerStarted","Data":"f7765dcc6e6ead6c26720c5dfa0b89bbac4f7e0540d818440814c45bec737b89"} Dec 09 11:14:36 crc kubenswrapper[4824]: I1209 11:14:36.950138 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-2cbmd" podUID="257d6d77-75f1-41e0-8bd2-c6ec3c1ff582" containerName="registry-server" probeResult="failure" output=< Dec 09 11:14:36 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 11:14:36 crc kubenswrapper[4824]: > Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.339874 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.402199 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0fd5955c-4039-42a1-8ecd-0256bb683e61-openstack-config-secret\") pod \"0fd5955c-4039-42a1-8ecd-0256bb683e61\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.402300 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzfmv\" (UniqueName: \"kubernetes.io/projected/0fd5955c-4039-42a1-8ecd-0256bb683e61-kube-api-access-dzfmv\") pod \"0fd5955c-4039-42a1-8ecd-0256bb683e61\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.402334 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0fd5955c-4039-42a1-8ecd-0256bb683e61-openstack-config\") pod \"0fd5955c-4039-42a1-8ecd-0256bb683e61\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.402388 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0fd5955c-4039-42a1-8ecd-0256bb683e61-ca-certs\") pod \"0fd5955c-4039-42a1-8ecd-0256bb683e61\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.402421 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fd5955c-4039-42a1-8ecd-0256bb683e61-ssh-key\") pod \"0fd5955c-4039-42a1-8ecd-0256bb683e61\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.402483 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"0fd5955c-4039-42a1-8ecd-0256bb683e61\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.402641 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0fd5955c-4039-42a1-8ecd-0256bb683e61-config-data\") pod \"0fd5955c-4039-42a1-8ecd-0256bb683e61\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.402835 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0fd5955c-4039-42a1-8ecd-0256bb683e61-test-operator-ephemeral-workdir\") pod \"0fd5955c-4039-42a1-8ecd-0256bb683e61\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.402869 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0fd5955c-4039-42a1-8ecd-0256bb683e61-test-operator-ephemeral-temporary\") pod \"0fd5955c-4039-42a1-8ecd-0256bb683e61\" (UID: \"0fd5955c-4039-42a1-8ecd-0256bb683e61\") " Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.405285 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fd5955c-4039-42a1-8ecd-0256bb683e61-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "0fd5955c-4039-42a1-8ecd-0256bb683e61" (UID: "0fd5955c-4039-42a1-8ecd-0256bb683e61"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.405521 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fd5955c-4039-42a1-8ecd-0256bb683e61-config-data" (OuterVolumeSpecName: "config-data") pod "0fd5955c-4039-42a1-8ecd-0256bb683e61" (UID: "0fd5955c-4039-42a1-8ecd-0256bb683e61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.410859 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fd5955c-4039-42a1-8ecd-0256bb683e61-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "0fd5955c-4039-42a1-8ecd-0256bb683e61" (UID: "0fd5955c-4039-42a1-8ecd-0256bb683e61"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.428616 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fd5955c-4039-42a1-8ecd-0256bb683e61-kube-api-access-dzfmv" (OuterVolumeSpecName: "kube-api-access-dzfmv") pod "0fd5955c-4039-42a1-8ecd-0256bb683e61" (UID: "0fd5955c-4039-42a1-8ecd-0256bb683e61"). InnerVolumeSpecName "kube-api-access-dzfmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.429050 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "test-operator-logs") pod "0fd5955c-4039-42a1-8ecd-0256bb683e61" (UID: "0fd5955c-4039-42a1-8ecd-0256bb683e61"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.457197 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fd5955c-4039-42a1-8ecd-0256bb683e61-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0fd5955c-4039-42a1-8ecd-0256bb683e61" (UID: "0fd5955c-4039-42a1-8ecd-0256bb683e61"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.462130 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fd5955c-4039-42a1-8ecd-0256bb683e61-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "0fd5955c-4039-42a1-8ecd-0256bb683e61" (UID: "0fd5955c-4039-42a1-8ecd-0256bb683e61"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.465461 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fd5955c-4039-42a1-8ecd-0256bb683e61-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "0fd5955c-4039-42a1-8ecd-0256bb683e61" (UID: "0fd5955c-4039-42a1-8ecd-0256bb683e61"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.500903 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fd5955c-4039-42a1-8ecd-0256bb683e61-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "0fd5955c-4039-42a1-8ecd-0256bb683e61" (UID: "0fd5955c-4039-42a1-8ecd-0256bb683e61"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.513937 4824 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0fd5955c-4039-42a1-8ecd-0256bb683e61-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.514140 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fd5955c-4039-42a1-8ecd-0256bb683e61-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.514465 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.514585 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0fd5955c-4039-42a1-8ecd-0256bb683e61-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.514672 4824 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0fd5955c-4039-42a1-8ecd-0256bb683e61-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.514740 4824 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0fd5955c-4039-42a1-8ecd-0256bb683e61-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.514815 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0fd5955c-4039-42a1-8ecd-0256bb683e61-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.514890 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzfmv\" (UniqueName: \"kubernetes.io/projected/0fd5955c-4039-42a1-8ecd-0256bb683e61-kube-api-access-dzfmv\") on node \"crc\" DevicePath \"\"" Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.514979 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0fd5955c-4039-42a1-8ecd-0256bb683e61-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.542575 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"0fd5955c-4039-42a1-8ecd-0256bb683e61","Type":"ContainerDied","Data":"9f448c061e67b0ec27df28f2f694dbdbdce723f631d1c133ba2464a33b6531c5"} Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.542636 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.542937 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f448c061e67b0ec27df28f2f694dbdbdce723f631d1c133ba2464a33b6531c5" Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.555416 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.564511 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879ftr284" Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.617549 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.744734 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 09 11:14:37 crc kubenswrapper[4824]: I1209 11:14:37.744835 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 09 11:14:38 crc kubenswrapper[4824]: I1209 11:14:38.311293 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7f84c5b8f9-rqmr9" Dec 09 11:14:38 crc kubenswrapper[4824]: E1209 11:14:38.516867 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfeab57e6_788b_4f23_9e99_aa248786052d.slice/crio-fe509de64aa6046a4afdd744d7b9a144392cce3237541624889199bb17b79b7b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee9b1e8c_b0c9_47ee_8c98_e8fe5153b6bc.slice/crio-conmon-8ef0f4d72f39e4c91f5d814ad0f5248f69fc7641a71cf06a32de5125e7d45dc5.scope\": RecentStats: unable to find data in memory cache]" Dec 09 11:14:39 crc kubenswrapper[4824]: I1209 11:14:39.349403 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 09 11:14:39 crc kubenswrapper[4824]: I1209 11:14:39.349757 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 09 11:14:39 crc kubenswrapper[4824]: I1209 11:14:39.811293 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5dcb7584cb-dfrlw" Dec 09 11:14:40 crc kubenswrapper[4824]: I1209 11:14:40.836166 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-wxgt4" Dec 09 11:14:41 crc kubenswrapper[4824]: I1209 11:14:41.910519 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:14:42 crc kubenswrapper[4824]: E1209 11:14:42.051177 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 11:14:42 crc kubenswrapper[4824]: E1209 11:14:42.051246 4824 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 11:14:42 crc kubenswrapper[4824]: E1209 11:14:42.051466 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n9h55h68ch65bh655h64chbfh6ch5b4h575h6bh654h5bdh587h5h58bh5dch5b9hf4h548h579hf8h5f5h686h667h97h57hb5h65h656h588h5f8q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6p7fx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9339c10c-1cea-4ffc-8931-9b349a1e74cd): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 11:14:42 crc kubenswrapper[4824]: E1209 11:14:42.052751 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:14:42 crc kubenswrapper[4824]: E1209 11:14:42.203599 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfeab57e6_788b_4f23_9e99_aa248786052d.slice/crio-fe509de64aa6046a4afdd744d7b9a144392cce3237541624889199bb17b79b7b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee9b1e8c_b0c9_47ee_8c98_e8fe5153b6bc.slice/crio-conmon-8ef0f4d72f39e4c91f5d814ad0f5248f69fc7641a71cf06a32de5125e7d45dc5.scope\": RecentStats: unable to find data in memory cache]" Dec 09 11:14:42 crc kubenswrapper[4824]: I1209 11:14:42.286976 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-5xfv2" Dec 09 11:14:42 crc kubenswrapper[4824]: I1209 11:14:42.599696 4824 generic.go:334] "Generic (PLEG): container finished" podID="153fe73c-2767-4f80-a7b4-0af49257065a" containerID="d2f14825848e7e355909e904cb07d3ac37df2eec491e11d6f97757df6a0f4a7d" exitCode=0 Dec 09 11:14:42 crc kubenswrapper[4824]: I1209 11:14:42.599744 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"153fe73c-2767-4f80-a7b4-0af49257065a","Type":"ContainerDied","Data":"d2f14825848e7e355909e904cb07d3ac37df2eec491e11d6f97757df6a0f4a7d"} Dec 09 11:14:42 crc kubenswrapper[4824]: I1209 11:14:42.646019 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" podUID="3d1e2cd1-5fc5-4a76-b1ff-07db11b07708" containerName="oauth-openshift" containerID="cri-o://166928d0068d6f77eb364346c00c3dd78dbb7602e7badaa6af86f3ce7f5c0df1" gracePeriod=13 Dec 09 11:14:43 crc kubenswrapper[4824]: I1209 11:14:43.616551 4824 generic.go:334] "Generic (PLEG): container finished" podID="3d1e2cd1-5fc5-4a76-b1ff-07db11b07708" containerID="166928d0068d6f77eb364346c00c3dd78dbb7602e7badaa6af86f3ce7f5c0df1" exitCode=0 Dec 09 11:14:43 crc kubenswrapper[4824]: I1209 11:14:43.616662 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" event={"ID":"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708","Type":"ContainerDied","Data":"166928d0068d6f77eb364346c00c3dd78dbb7602e7badaa6af86f3ce7f5c0df1"} Dec 09 11:14:43 crc kubenswrapper[4824]: I1209 11:14:43.617684 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" event={"ID":"3d1e2cd1-5fc5-4a76-b1ff-07db11b07708","Type":"ContainerStarted","Data":"bfb62c518b195d60065de8d0de8130da504223527964132874cdbcf4ec3d1127"} Dec 09 11:14:43 crc kubenswrapper[4824]: I1209 11:14:43.617701 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 11:14:43 crc kubenswrapper[4824]: I1209 11:14:43.621246 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"153fe73c-2767-4f80-a7b4-0af49257065a","Type":"ContainerStarted","Data":"45a620d49647084db738dbe84f4d6bdc4df0ea10ae4b3d9688b59a6d4bab7d0b"} Dec 09 11:14:43 crc kubenswrapper[4824]: I1209 11:14:43.624477 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-f578d5c8f-5jzzv" Dec 09 11:14:43 crc kubenswrapper[4824]: I1209 11:14:43.628292 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-95jfr" Dec 09 11:14:43 crc kubenswrapper[4824]: I1209 11:14:43.720321 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-95jfr" Dec 09 11:14:44 crc kubenswrapper[4824]: I1209 11:14:44.290585 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-582xj" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" containerName="registry-server" probeResult="failure" output=< Dec 09 11:14:44 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 11:14:44 crc kubenswrapper[4824]: > Dec 09 11:14:45 crc kubenswrapper[4824]: I1209 11:14:45.368320 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jm5ld" Dec 09 11:14:45 crc kubenswrapper[4824]: I1209 11:14:45.433408 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jm5ld" Dec 09 11:14:45 crc kubenswrapper[4824]: I1209 11:14:45.957718 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2cbmd" Dec 09 11:14:46 crc kubenswrapper[4824]: I1209 11:14:46.012293 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2cbmd" Dec 09 11:14:46 crc kubenswrapper[4824]: I1209 11:14:46.611550 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 09 11:14:46 crc kubenswrapper[4824]: E1209 11:14:46.612380 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91c1710e-cd28-4849-b232-9af634a3b317" containerName="registry-server" Dec 09 11:14:46 crc kubenswrapper[4824]: I1209 11:14:46.612397 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="91c1710e-cd28-4849-b232-9af634a3b317" containerName="registry-server" Dec 09 11:14:46 crc kubenswrapper[4824]: E1209 11:14:46.612411 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cd5d224-1a1e-4154-9b44-4c6b0e9ea657" containerName="extract-content" Dec 09 11:14:46 crc kubenswrapper[4824]: I1209 11:14:46.612418 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cd5d224-1a1e-4154-9b44-4c6b0e9ea657" containerName="extract-content" Dec 09 11:14:46 crc kubenswrapper[4824]: E1209 11:14:46.612439 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cd5d224-1a1e-4154-9b44-4c6b0e9ea657" containerName="registry-server" Dec 09 11:14:46 crc kubenswrapper[4824]: I1209 11:14:46.612445 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cd5d224-1a1e-4154-9b44-4c6b0e9ea657" containerName="registry-server" Dec 09 11:14:46 crc kubenswrapper[4824]: E1209 11:14:46.612457 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fd5955c-4039-42a1-8ecd-0256bb683e61" containerName="tempest-tests-tempest-tests-runner" Dec 09 11:14:46 crc kubenswrapper[4824]: I1209 11:14:46.612463 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fd5955c-4039-42a1-8ecd-0256bb683e61" containerName="tempest-tests-tempest-tests-runner" Dec 09 11:14:46 crc kubenswrapper[4824]: E1209 11:14:46.612487 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91c1710e-cd28-4849-b232-9af634a3b317" containerName="extract-content" Dec 09 11:14:46 crc kubenswrapper[4824]: I1209 11:14:46.612493 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="91c1710e-cd28-4849-b232-9af634a3b317" containerName="extract-content" Dec 09 11:14:46 crc kubenswrapper[4824]: E1209 11:14:46.612507 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91c1710e-cd28-4849-b232-9af634a3b317" containerName="extract-utilities" Dec 09 11:14:46 crc kubenswrapper[4824]: I1209 11:14:46.612513 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="91c1710e-cd28-4849-b232-9af634a3b317" containerName="extract-utilities" Dec 09 11:14:46 crc kubenswrapper[4824]: E1209 11:14:46.612524 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cd5d224-1a1e-4154-9b44-4c6b0e9ea657" containerName="extract-utilities" Dec 09 11:14:46 crc kubenswrapper[4824]: I1209 11:14:46.612530 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cd5d224-1a1e-4154-9b44-4c6b0e9ea657" containerName="extract-utilities" Dec 09 11:14:46 crc kubenswrapper[4824]: I1209 11:14:46.612768 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cd5d224-1a1e-4154-9b44-4c6b0e9ea657" containerName="registry-server" Dec 09 11:14:46 crc kubenswrapper[4824]: I1209 11:14:46.612802 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fd5955c-4039-42a1-8ecd-0256bb683e61" containerName="tempest-tests-tempest-tests-runner" Dec 09 11:14:46 crc kubenswrapper[4824]: I1209 11:14:46.612827 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="91c1710e-cd28-4849-b232-9af634a3b317" containerName="registry-server" Dec 09 11:14:46 crc kubenswrapper[4824]: I1209 11:14:46.614393 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 09 11:14:46 crc kubenswrapper[4824]: I1209 11:14:46.617063 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-n76qc" Dec 09 11:14:46 crc kubenswrapper[4824]: I1209 11:14:46.627194 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 09 11:14:46 crc kubenswrapper[4824]: I1209 11:14:46.757123 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d29986ea-5ca9-442e-999b-48ac99ddcc9e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 09 11:14:46 crc kubenswrapper[4824]: I1209 11:14:46.757735 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfh8j\" (UniqueName: \"kubernetes.io/projected/d29986ea-5ca9-442e-999b-48ac99ddcc9e-kube-api-access-vfh8j\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d29986ea-5ca9-442e-999b-48ac99ddcc9e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 09 11:14:46 crc kubenswrapper[4824]: I1209 11:14:46.860586 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d29986ea-5ca9-442e-999b-48ac99ddcc9e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 09 11:14:46 crc kubenswrapper[4824]: I1209 11:14:46.860700 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfh8j\" (UniqueName: \"kubernetes.io/projected/d29986ea-5ca9-442e-999b-48ac99ddcc9e-kube-api-access-vfh8j\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d29986ea-5ca9-442e-999b-48ac99ddcc9e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 09 11:14:46 crc kubenswrapper[4824]: I1209 11:14:46.861160 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d29986ea-5ca9-442e-999b-48ac99ddcc9e\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 09 11:14:46 crc kubenswrapper[4824]: I1209 11:14:46.924241 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfh8j\" (UniqueName: \"kubernetes.io/projected/d29986ea-5ca9-442e-999b-48ac99ddcc9e-kube-api-access-vfh8j\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d29986ea-5ca9-442e-999b-48ac99ddcc9e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 09 11:14:46 crc kubenswrapper[4824]: I1209 11:14:46.934302 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d29986ea-5ca9-442e-999b-48ac99ddcc9e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 09 11:14:47 crc kubenswrapper[4824]: I1209 11:14:47.236263 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 09 11:14:47 crc kubenswrapper[4824]: I1209 11:14:47.896080 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 09 11:14:48 crc kubenswrapper[4824]: I1209 11:14:48.008655 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 09 11:14:48 crc kubenswrapper[4824]: W1209 11:14:48.011571 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd29986ea_5ca9_442e_999b_48ac99ddcc9e.slice/crio-5ece58be42fb667868856e81e8dbcb4172d81902de6f97dbbd50931182c8468f WatchSource:0}: Error finding container 5ece58be42fb667868856e81e8dbcb4172d81902de6f97dbbd50931182c8468f: Status 404 returned error can't find the container with id 5ece58be42fb667868856e81e8dbcb4172d81902de6f97dbbd50931182c8468f Dec 09 11:14:48 crc kubenswrapper[4824]: E1209 11:14:48.255319 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee9b1e8c_b0c9_47ee_8c98_e8fe5153b6bc.slice/crio-conmon-8ef0f4d72f39e4c91f5d814ad0f5248f69fc7641a71cf06a32de5125e7d45dc5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfeab57e6_788b_4f23_9e99_aa248786052d.slice/crio-fe509de64aa6046a4afdd744d7b9a144392cce3237541624889199bb17b79b7b.scope\": RecentStats: unable to find data in memory cache]" Dec 09 11:14:48 crc kubenswrapper[4824]: E1209 11:14:48.255730 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfeab57e6_788b_4f23_9e99_aa248786052d.slice/crio-fe509de64aa6046a4afdd744d7b9a144392cce3237541624889199bb17b79b7b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee9b1e8c_b0c9_47ee_8c98_e8fe5153b6bc.slice/crio-conmon-8ef0f4d72f39e4c91f5d814ad0f5248f69fc7641a71cf06a32de5125e7d45dc5.scope\": RecentStats: unable to find data in memory cache]" Dec 09 11:14:48 crc kubenswrapper[4824]: I1209 11:14:48.682870 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"d29986ea-5ca9-442e-999b-48ac99ddcc9e","Type":"ContainerStarted","Data":"5ece58be42fb667868856e81e8dbcb4172d81902de6f97dbbd50931182c8468f"} Dec 09 11:14:52 crc kubenswrapper[4824]: E1209 11:14:52.250378 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee9b1e8c_b0c9_47ee_8c98_e8fe5153b6bc.slice/crio-conmon-8ef0f4d72f39e4c91f5d814ad0f5248f69fc7641a71cf06a32de5125e7d45dc5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfeab57e6_788b_4f23_9e99_aa248786052d.slice/crio-fe509de64aa6046a4afdd744d7b9a144392cce3237541624889199bb17b79b7b.scope\": RecentStats: unable to find data in memory cache]" Dec 09 11:14:52 crc kubenswrapper[4824]: I1209 11:14:52.918729 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="153fe73c-2767-4f80-a7b4-0af49257065a" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:14:53 crc kubenswrapper[4824]: E1209 11:14:53.620855 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee9b1e8c_b0c9_47ee_8c98_e8fe5153b6bc.slice/crio-conmon-8ef0f4d72f39e4c91f5d814ad0f5248f69fc7641a71cf06a32de5125e7d45dc5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfeab57e6_788b_4f23_9e99_aa248786052d.slice/crio-fe509de64aa6046a4afdd744d7b9a144392cce3237541624889199bb17b79b7b.scope\": RecentStats: unable to find data in memory cache]" Dec 09 11:14:54 crc kubenswrapper[4824]: I1209 11:14:54.511215 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-582xj" podUID="f2572cf3-89f1-49d7-9246-9350703e9764" containerName="registry-server" probeResult="failure" output=< Dec 09 11:14:54 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 11:14:54 crc kubenswrapper[4824]: > Dec 09 11:14:54 crc kubenswrapper[4824]: I1209 11:14:54.769923 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"d29986ea-5ca9-442e-999b-48ac99ddcc9e","Type":"ContainerStarted","Data":"121fad5bcb15ca7c51be661618df2741075b3e6f1d8d66172567cec30d65600b"} Dec 09 11:14:55 crc kubenswrapper[4824]: I1209 11:14:55.264260 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.763964949 podStartE2EDuration="8.807940351s" podCreationTimestamp="2025-12-09 11:14:46 +0000 UTC" firstStartedPulling="2025-12-09 11:14:48.013406172 +0000 UTC m=+5244.347910839" lastFinishedPulling="2025-12-09 11:14:54.057381574 +0000 UTC m=+5250.391886241" observedRunningTime="2025-12-09 11:14:54.782347159 +0000 UTC m=+5251.116851826" watchObservedRunningTime="2025-12-09 11:14:54.807940351 +0000 UTC m=+5251.142445018" Dec 09 11:14:55 crc kubenswrapper[4824]: I1209 11:14:55.911348 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:14:55 crc kubenswrapper[4824]: E1209 11:14:55.913419 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:14:57 crc kubenswrapper[4824]: I1209 11:14:57.918188 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="153fe73c-2767-4f80-a7b4-0af49257065a" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:15:00 crc kubenswrapper[4824]: I1209 11:15:00.190998 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421315-msd4n"] Dec 09 11:15:00 crc kubenswrapper[4824]: I1209 11:15:00.193852 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421315-msd4n" Dec 09 11:15:00 crc kubenswrapper[4824]: I1209 11:15:00.197706 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 11:15:00 crc kubenswrapper[4824]: I1209 11:15:00.197825 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 11:15:00 crc kubenswrapper[4824]: I1209 11:15:00.207631 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421315-msd4n"] Dec 09 11:15:00 crc kubenswrapper[4824]: I1209 11:15:00.222167 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dd73ed78-c3a1-421d-8900-95d89783dbab-config-volume\") pod \"collect-profiles-29421315-msd4n\" (UID: \"dd73ed78-c3a1-421d-8900-95d89783dbab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421315-msd4n" Dec 09 11:15:00 crc kubenswrapper[4824]: I1209 11:15:00.222236 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dd73ed78-c3a1-421d-8900-95d89783dbab-secret-volume\") pod \"collect-profiles-29421315-msd4n\" (UID: \"dd73ed78-c3a1-421d-8900-95d89783dbab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421315-msd4n" Dec 09 11:15:00 crc kubenswrapper[4824]: I1209 11:15:00.222328 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbmch\" (UniqueName: \"kubernetes.io/projected/dd73ed78-c3a1-421d-8900-95d89783dbab-kube-api-access-mbmch\") pod \"collect-profiles-29421315-msd4n\" (UID: \"dd73ed78-c3a1-421d-8900-95d89783dbab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421315-msd4n" Dec 09 11:15:00 crc kubenswrapper[4824]: I1209 11:15:00.325197 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dd73ed78-c3a1-421d-8900-95d89783dbab-config-volume\") pod \"collect-profiles-29421315-msd4n\" (UID: \"dd73ed78-c3a1-421d-8900-95d89783dbab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421315-msd4n" Dec 09 11:15:00 crc kubenswrapper[4824]: I1209 11:15:00.325262 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dd73ed78-c3a1-421d-8900-95d89783dbab-secret-volume\") pod \"collect-profiles-29421315-msd4n\" (UID: \"dd73ed78-c3a1-421d-8900-95d89783dbab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421315-msd4n" Dec 09 11:15:00 crc kubenswrapper[4824]: I1209 11:15:00.325455 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbmch\" (UniqueName: \"kubernetes.io/projected/dd73ed78-c3a1-421d-8900-95d89783dbab-kube-api-access-mbmch\") pod \"collect-profiles-29421315-msd4n\" (UID: \"dd73ed78-c3a1-421d-8900-95d89783dbab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421315-msd4n" Dec 09 11:15:00 crc kubenswrapper[4824]: I1209 11:15:00.327586 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dd73ed78-c3a1-421d-8900-95d89783dbab-config-volume\") pod \"collect-profiles-29421315-msd4n\" (UID: \"dd73ed78-c3a1-421d-8900-95d89783dbab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421315-msd4n" Dec 09 11:15:00 crc kubenswrapper[4824]: I1209 11:15:00.332890 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dd73ed78-c3a1-421d-8900-95d89783dbab-secret-volume\") pod \"collect-profiles-29421315-msd4n\" (UID: \"dd73ed78-c3a1-421d-8900-95d89783dbab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421315-msd4n" Dec 09 11:15:00 crc kubenswrapper[4824]: I1209 11:15:00.347983 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbmch\" (UniqueName: \"kubernetes.io/projected/dd73ed78-c3a1-421d-8900-95d89783dbab-kube-api-access-mbmch\") pod \"collect-profiles-29421315-msd4n\" (UID: \"dd73ed78-c3a1-421d-8900-95d89783dbab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421315-msd4n" Dec 09 11:15:00 crc kubenswrapper[4824]: I1209 11:15:00.528660 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421315-msd4n" Dec 09 11:15:01 crc kubenswrapper[4824]: I1209 11:15:01.102494 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421315-msd4n"] Dec 09 11:15:01 crc kubenswrapper[4824]: W1209 11:15:01.120738 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd73ed78_c3a1_421d_8900_95d89783dbab.slice/crio-d21ca27c8c4b9fc82de1deb23b398ddbf317d1fa906f8f76f04105594d16a4fa WatchSource:0}: Error finding container d21ca27c8c4b9fc82de1deb23b398ddbf317d1fa906f8f76f04105594d16a4fa: Status 404 returned error can't find the container with id d21ca27c8c4b9fc82de1deb23b398ddbf317d1fa906f8f76f04105594d16a4fa Dec 09 11:15:01 crc kubenswrapper[4824]: I1209 11:15:01.847767 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421315-msd4n" event={"ID":"dd73ed78-c3a1-421d-8900-95d89783dbab","Type":"ContainerStarted","Data":"d21ca27c8c4b9fc82de1deb23b398ddbf317d1fa906f8f76f04105594d16a4fa"} Dec 09 11:15:02 crc kubenswrapper[4824]: E1209 11:15:02.597668 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee9b1e8c_b0c9_47ee_8c98_e8fe5153b6bc.slice/crio-conmon-8ef0f4d72f39e4c91f5d814ad0f5248f69fc7641a71cf06a32de5125e7d45dc5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfeab57e6_788b_4f23_9e99_aa248786052d.slice/crio-fe509de64aa6046a4afdd744d7b9a144392cce3237541624889199bb17b79b7b.scope\": RecentStats: unable to find data in memory cache]" Dec 09 11:15:02 crc kubenswrapper[4824]: I1209 11:15:02.862510 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:15:02 crc kubenswrapper[4824]: I1209 11:15:02.862577 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:15:03 crc kubenswrapper[4824]: I1209 11:15:02.865168 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421315-msd4n" event={"ID":"dd73ed78-c3a1-421d-8900-95d89783dbab","Type":"ContainerStarted","Data":"c81678797f9f109ac0be7f6baee4aa2dfca322cb252b41256ad6bcf219b92171"} Dec 09 11:15:03 crc kubenswrapper[4824]: I1209 11:15:02.935197 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="153fe73c-2767-4f80-a7b4-0af49257065a" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:15:03 crc kubenswrapper[4824]: I1209 11:15:03.983237 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-582xj" Dec 09 11:15:04 crc kubenswrapper[4824]: I1209 11:15:04.060336 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-582xj" Dec 09 11:15:04 crc kubenswrapper[4824]: I1209 11:15:04.065316 4824 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-k8bgl container/package-server-manager namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:15:04 crc kubenswrapper[4824]: I1209 11:15:04.065410 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" podUID="6c10edf9-cf75-4633-8d4c-9bbb9d9b6339" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:15:04 crc kubenswrapper[4824]: I1209 11:15:04.066485 4824 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-k8bgl container/package-server-manager namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 11:15:04 crc kubenswrapper[4824]: I1209 11:15:04.066523 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" podUID="6c10edf9-cf75-4633-8d4c-9bbb9d9b6339" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 11:15:04 crc kubenswrapper[4824]: I1209 11:15:04.897930 4824 generic.go:334] "Generic (PLEG): container finished" podID="dd73ed78-c3a1-421d-8900-95d89783dbab" containerID="c81678797f9f109ac0be7f6baee4aa2dfca322cb252b41256ad6bcf219b92171" exitCode=0 Dec 09 11:15:04 crc kubenswrapper[4824]: I1209 11:15:04.899986 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421315-msd4n" event={"ID":"dd73ed78-c3a1-421d-8900-95d89783dbab","Type":"ContainerDied","Data":"c81678797f9f109ac0be7f6baee4aa2dfca322cb252b41256ad6bcf219b92171"} Dec 09 11:15:06 crc kubenswrapper[4824]: I1209 11:15:06.568425 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421315-msd4n" Dec 09 11:15:06 crc kubenswrapper[4824]: I1209 11:15:06.639012 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dd73ed78-c3a1-421d-8900-95d89783dbab-config-volume\") pod \"dd73ed78-c3a1-421d-8900-95d89783dbab\" (UID: \"dd73ed78-c3a1-421d-8900-95d89783dbab\") " Dec 09 11:15:06 crc kubenswrapper[4824]: I1209 11:15:06.639155 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbmch\" (UniqueName: \"kubernetes.io/projected/dd73ed78-c3a1-421d-8900-95d89783dbab-kube-api-access-mbmch\") pod \"dd73ed78-c3a1-421d-8900-95d89783dbab\" (UID: \"dd73ed78-c3a1-421d-8900-95d89783dbab\") " Dec 09 11:15:06 crc kubenswrapper[4824]: I1209 11:15:06.639228 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dd73ed78-c3a1-421d-8900-95d89783dbab-secret-volume\") pod \"dd73ed78-c3a1-421d-8900-95d89783dbab\" (UID: \"dd73ed78-c3a1-421d-8900-95d89783dbab\") " Dec 09 11:15:06 crc kubenswrapper[4824]: I1209 11:15:06.641571 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd73ed78-c3a1-421d-8900-95d89783dbab-config-volume" (OuterVolumeSpecName: "config-volume") pod "dd73ed78-c3a1-421d-8900-95d89783dbab" (UID: "dd73ed78-c3a1-421d-8900-95d89783dbab"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 11:15:06 crc kubenswrapper[4824]: I1209 11:15:06.649186 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd73ed78-c3a1-421d-8900-95d89783dbab-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "dd73ed78-c3a1-421d-8900-95d89783dbab" (UID: "dd73ed78-c3a1-421d-8900-95d89783dbab"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 11:15:06 crc kubenswrapper[4824]: I1209 11:15:06.649435 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd73ed78-c3a1-421d-8900-95d89783dbab-kube-api-access-mbmch" (OuterVolumeSpecName: "kube-api-access-mbmch") pod "dd73ed78-c3a1-421d-8900-95d89783dbab" (UID: "dd73ed78-c3a1-421d-8900-95d89783dbab"). InnerVolumeSpecName "kube-api-access-mbmch". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:15:06 crc kubenswrapper[4824]: I1209 11:15:06.745391 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dd73ed78-c3a1-421d-8900-95d89783dbab-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 11:15:06 crc kubenswrapper[4824]: I1209 11:15:06.745433 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbmch\" (UniqueName: \"kubernetes.io/projected/dd73ed78-c3a1-421d-8900-95d89783dbab-kube-api-access-mbmch\") on node \"crc\" DevicePath \"\"" Dec 09 11:15:06 crc kubenswrapper[4824]: I1209 11:15:06.745447 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dd73ed78-c3a1-421d-8900-95d89783dbab-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 11:15:06 crc kubenswrapper[4824]: I1209 11:15:06.925840 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421315-msd4n" event={"ID":"dd73ed78-c3a1-421d-8900-95d89783dbab","Type":"ContainerDied","Data":"d21ca27c8c4b9fc82de1deb23b398ddbf317d1fa906f8f76f04105594d16a4fa"} Dec 09 11:15:06 crc kubenswrapper[4824]: I1209 11:15:06.925885 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d21ca27c8c4b9fc82de1deb23b398ddbf317d1fa906f8f76f04105594d16a4fa" Dec 09 11:15:06 crc kubenswrapper[4824]: I1209 11:15:06.925893 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421315-msd4n" Dec 09 11:15:07 crc kubenswrapper[4824]: I1209 11:15:07.743546 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421270-8wp4p"] Dec 09 11:15:07 crc kubenswrapper[4824]: I1209 11:15:07.767652 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421270-8wp4p"] Dec 09 11:15:07 crc kubenswrapper[4824]: I1209 11:15:07.922373 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="153fe73c-2767-4f80-a7b4-0af49257065a" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 11:15:07 crc kubenswrapper[4824]: I1209 11:15:07.987626 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65d6395b-1cca-479b-986c-49b2142e7a30" path="/var/lib/kubelet/pods/65d6395b-1cca-479b-986c-49b2142e7a30/volumes" Dec 09 11:15:08 crc kubenswrapper[4824]: I1209 11:15:08.535532 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 09 11:15:08 crc kubenswrapper[4824]: E1209 11:15:08.609719 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee9b1e8c_b0c9_47ee_8c98_e8fe5153b6bc.slice/crio-conmon-8ef0f4d72f39e4c91f5d814ad0f5248f69fc7641a71cf06a32de5125e7d45dc5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfeab57e6_788b_4f23_9e99_aa248786052d.slice/crio-fe509de64aa6046a4afdd744d7b9a144392cce3237541624889199bb17b79b7b.scope\": RecentStats: unable to find data in memory cache]" Dec 09 11:15:08 crc kubenswrapper[4824]: I1209 11:15:08.671283 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 09 11:15:09 crc kubenswrapper[4824]: I1209 11:15:09.911589 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:15:10 crc kubenswrapper[4824]: E1209 11:15:10.039236 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 11:15:10 crc kubenswrapper[4824]: E1209 11:15:10.039439 4824 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 11:15:10 crc kubenswrapper[4824]: E1209 11:15:10.039963 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n9h55h68ch65bh655h64chbfh6ch5b4h575h6bh654h5bdh587h5h58bh5dch5b9hf4h548h579hf8h5f5h686h667h97h57hb5h65h656h588h5f8q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6p7fx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9339c10c-1cea-4ffc-8931-9b349a1e74cd): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 11:15:10 crc kubenswrapper[4824]: E1209 11:15:10.041626 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:15:10 crc kubenswrapper[4824]: I1209 11:15:10.708411 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 09 11:15:10 crc kubenswrapper[4824]: I1209 11:15:10.814968 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 09 11:15:12 crc kubenswrapper[4824]: E1209 11:15:12.680165 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfeab57e6_788b_4f23_9e99_aa248786052d.slice/crio-fe509de64aa6046a4afdd744d7b9a144392cce3237541624889199bb17b79b7b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee9b1e8c_b0c9_47ee_8c98_e8fe5153b6bc.slice/crio-conmon-8ef0f4d72f39e4c91f5d814ad0f5248f69fc7641a71cf06a32de5125e7d45dc5.scope\": RecentStats: unable to find data in memory cache]" Dec 09 11:15:12 crc kubenswrapper[4824]: I1209 11:15:12.934993 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 09 11:15:12 crc kubenswrapper[4824]: I1209 11:15:12.968461 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-k8bgl" Dec 09 11:15:16 crc kubenswrapper[4824]: I1209 11:15:16.077677 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 11:15:20 crc kubenswrapper[4824]: I1209 11:15:20.637479 4824 scope.go:117] "RemoveContainer" containerID="91cc1a2de21bfb84fc95fdf995d849a2432267276a144e63a9c420151952f83e" Dec 09 11:15:23 crc kubenswrapper[4824]: E1209 11:15:23.068221 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfeab57e6_788b_4f23_9e99_aa248786052d.slice/crio-fe509de64aa6046a4afdd744d7b9a144392cce3237541624889199bb17b79b7b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee9b1e8c_b0c9_47ee_8c98_e8fe5153b6bc.slice/crio-conmon-8ef0f4d72f39e4c91f5d814ad0f5248f69fc7641a71cf06a32de5125e7d45dc5.scope\": RecentStats: unable to find data in memory cache]" Dec 09 11:15:23 crc kubenswrapper[4824]: E1209 11:15:23.249444 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfeab57e6_788b_4f23_9e99_aa248786052d.slice/crio-fe509de64aa6046a4afdd744d7b9a144392cce3237541624889199bb17b79b7b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee9b1e8c_b0c9_47ee_8c98_e8fe5153b6bc.slice/crio-conmon-8ef0f4d72f39e4c91f5d814ad0f5248f69fc7641a71cf06a32de5125e7d45dc5.scope\": RecentStats: unable to find data in memory cache]" Dec 09 11:15:23 crc kubenswrapper[4824]: I1209 11:15:23.934916 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:15:23 crc kubenswrapper[4824]: E1209 11:15:23.935335 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=ceilometer-central-agent pod=ceilometer-0_openstack(9339c10c-1cea-4ffc-8931-9b349a1e74cd)\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:15:32 crc kubenswrapper[4824]: I1209 11:15:32.861071 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:15:32 crc kubenswrapper[4824]: I1209 11:15:32.861668 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:15:36 crc kubenswrapper[4824]: I1209 11:15:36.910612 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:15:36 crc kubenswrapper[4824]: E1209 11:15:36.911441 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=ceilometer-central-agent pod=ceilometer-0_openstack(9339c10c-1cea-4ffc-8931-9b349a1e74cd)\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:15:45 crc kubenswrapper[4824]: I1209 11:15:45.410353 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2w2mq/must-gather-qd8qq"] Dec 09 11:15:45 crc kubenswrapper[4824]: E1209 11:15:45.412260 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd73ed78-c3a1-421d-8900-95d89783dbab" containerName="collect-profiles" Dec 09 11:15:45 crc kubenswrapper[4824]: I1209 11:15:45.412799 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd73ed78-c3a1-421d-8900-95d89783dbab" containerName="collect-profiles" Dec 09 11:15:45 crc kubenswrapper[4824]: I1209 11:15:45.413555 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd73ed78-c3a1-421d-8900-95d89783dbab" containerName="collect-profiles" Dec 09 11:15:45 crc kubenswrapper[4824]: I1209 11:15:45.416056 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2w2mq/must-gather-qd8qq" Dec 09 11:15:45 crc kubenswrapper[4824]: I1209 11:15:45.420443 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-2w2mq"/"default-dockercfg-qwh7z" Dec 09 11:15:45 crc kubenswrapper[4824]: I1209 11:15:45.420977 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-2w2mq"/"kube-root-ca.crt" Dec 09 11:15:45 crc kubenswrapper[4824]: I1209 11:15:45.426684 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-2w2mq"/"openshift-service-ca.crt" Dec 09 11:15:45 crc kubenswrapper[4824]: I1209 11:15:45.431420 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-2w2mq/must-gather-qd8qq"] Dec 09 11:15:45 crc kubenswrapper[4824]: I1209 11:15:45.509180 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzqbj\" (UniqueName: \"kubernetes.io/projected/594f6c5d-8833-4f51-bc27-52254f02ceaa-kube-api-access-qzqbj\") pod \"must-gather-qd8qq\" (UID: \"594f6c5d-8833-4f51-bc27-52254f02ceaa\") " pod="openshift-must-gather-2w2mq/must-gather-qd8qq" Dec 09 11:15:45 crc kubenswrapper[4824]: I1209 11:15:45.509839 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/594f6c5d-8833-4f51-bc27-52254f02ceaa-must-gather-output\") pod \"must-gather-qd8qq\" (UID: \"594f6c5d-8833-4f51-bc27-52254f02ceaa\") " pod="openshift-must-gather-2w2mq/must-gather-qd8qq" Dec 09 11:15:45 crc kubenswrapper[4824]: I1209 11:15:45.612099 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/594f6c5d-8833-4f51-bc27-52254f02ceaa-must-gather-output\") pod \"must-gather-qd8qq\" (UID: \"594f6c5d-8833-4f51-bc27-52254f02ceaa\") " pod="openshift-must-gather-2w2mq/must-gather-qd8qq" Dec 09 11:15:45 crc kubenswrapper[4824]: I1209 11:15:45.613029 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzqbj\" (UniqueName: \"kubernetes.io/projected/594f6c5d-8833-4f51-bc27-52254f02ceaa-kube-api-access-qzqbj\") pod \"must-gather-qd8qq\" (UID: \"594f6c5d-8833-4f51-bc27-52254f02ceaa\") " pod="openshift-must-gather-2w2mq/must-gather-qd8qq" Dec 09 11:15:45 crc kubenswrapper[4824]: I1209 11:15:45.612612 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/594f6c5d-8833-4f51-bc27-52254f02ceaa-must-gather-output\") pod \"must-gather-qd8qq\" (UID: \"594f6c5d-8833-4f51-bc27-52254f02ceaa\") " pod="openshift-must-gather-2w2mq/must-gather-qd8qq" Dec 09 11:15:46 crc kubenswrapper[4824]: I1209 11:15:46.053731 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzqbj\" (UniqueName: \"kubernetes.io/projected/594f6c5d-8833-4f51-bc27-52254f02ceaa-kube-api-access-qzqbj\") pod \"must-gather-qd8qq\" (UID: \"594f6c5d-8833-4f51-bc27-52254f02ceaa\") " pod="openshift-must-gather-2w2mq/must-gather-qd8qq" Dec 09 11:15:46 crc kubenswrapper[4824]: I1209 11:15:46.348958 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2w2mq/must-gather-qd8qq" Dec 09 11:15:47 crc kubenswrapper[4824]: I1209 11:15:47.019606 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-2w2mq/must-gather-qd8qq"] Dec 09 11:15:47 crc kubenswrapper[4824]: I1209 11:15:47.461497 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2w2mq/must-gather-qd8qq" event={"ID":"594f6c5d-8833-4f51-bc27-52254f02ceaa","Type":"ContainerStarted","Data":"11899ba62785f7f8ce887a5bb42cc4665c1e6417e60666f92ec22170463f3628"} Dec 09 11:15:50 crc kubenswrapper[4824]: I1209 11:15:50.911271 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:15:54 crc kubenswrapper[4824]: E1209 11:15:54.930549 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 11:15:54 crc kubenswrapper[4824]: E1209 11:15:54.931105 4824 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 11:15:54 crc kubenswrapper[4824]: E1209 11:15:54.931274 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n9h55h68ch65bh655h64chbfh6ch5b4h575h6bh654h5bdh587h5h58bh5dch5b9hf4h548h579hf8h5f5h686h667h97h57hb5h65h656h588h5f8q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6p7fx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9339c10c-1cea-4ffc-8931-9b349a1e74cd): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 11:15:54 crc kubenswrapper[4824]: E1209 11:15:54.937188 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:15:58 crc kubenswrapper[4824]: I1209 11:15:58.631615 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2w2mq/must-gather-qd8qq" event={"ID":"594f6c5d-8833-4f51-bc27-52254f02ceaa","Type":"ContainerStarted","Data":"65752f2c91d1d890d00554d049abfc70baa112c2f7e9d29aac6186ec3b03351d"} Dec 09 11:15:58 crc kubenswrapper[4824]: I1209 11:15:58.632194 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2w2mq/must-gather-qd8qq" event={"ID":"594f6c5d-8833-4f51-bc27-52254f02ceaa","Type":"ContainerStarted","Data":"2b92e298c4461b580890011a817668dce792070ce6e3fb399861eb96a5a8993d"} Dec 09 11:15:58 crc kubenswrapper[4824]: I1209 11:15:58.654743 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-2w2mq/must-gather-qd8qq" podStartSLOduration=2.832935606 podStartE2EDuration="13.654274639s" podCreationTimestamp="2025-12-09 11:15:45 +0000 UTC" firstStartedPulling="2025-12-09 11:15:47.038793854 +0000 UTC m=+5303.373298521" lastFinishedPulling="2025-12-09 11:15:57.860132847 +0000 UTC m=+5314.194637554" observedRunningTime="2025-12-09 11:15:58.650824912 +0000 UTC m=+5314.985329589" watchObservedRunningTime="2025-12-09 11:15:58.654274639 +0000 UTC m=+5314.988779326" Dec 09 11:16:02 crc kubenswrapper[4824]: I1209 11:16:02.867955 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:16:02 crc kubenswrapper[4824]: I1209 11:16:02.868738 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:16:02 crc kubenswrapper[4824]: I1209 11:16:02.869124 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 11:16:02 crc kubenswrapper[4824]: I1209 11:16:02.874254 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20"} pod="openshift-machine-config-operator/machine-config-daemon-dth8x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 11:16:02 crc kubenswrapper[4824]: I1209 11:16:02.877813 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" containerID="cri-o://84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" gracePeriod=600 Dec 09 11:16:03 crc kubenswrapper[4824]: E1209 11:16:03.005166 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:16:03 crc kubenswrapper[4824]: I1209 11:16:03.918265 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" exitCode=0 Dec 09 11:16:03 crc kubenswrapper[4824]: I1209 11:16:03.934708 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerDied","Data":"84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20"} Dec 09 11:16:03 crc kubenswrapper[4824]: I1209 11:16:03.934770 4824 scope.go:117] "RemoveContainer" containerID="a341fa1c8d0134458a190647787e775a46a7ce784fbe98c35ff532671dc5ea08" Dec 09 11:16:03 crc kubenswrapper[4824]: I1209 11:16:03.935840 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:16:03 crc kubenswrapper[4824]: E1209 11:16:03.936178 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:16:04 crc kubenswrapper[4824]: I1209 11:16:04.577434 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2w2mq/crc-debug-pb5xg"] Dec 09 11:16:04 crc kubenswrapper[4824]: I1209 11:16:04.580086 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2w2mq/crc-debug-pb5xg" Dec 09 11:16:04 crc kubenswrapper[4824]: I1209 11:16:04.617998 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/925ea289-5f58-4051-ab56-480312f713ef-host\") pod \"crc-debug-pb5xg\" (UID: \"925ea289-5f58-4051-ab56-480312f713ef\") " pod="openshift-must-gather-2w2mq/crc-debug-pb5xg" Dec 09 11:16:04 crc kubenswrapper[4824]: I1209 11:16:04.618339 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hkpn\" (UniqueName: \"kubernetes.io/projected/925ea289-5f58-4051-ab56-480312f713ef-kube-api-access-2hkpn\") pod \"crc-debug-pb5xg\" (UID: \"925ea289-5f58-4051-ab56-480312f713ef\") " pod="openshift-must-gather-2w2mq/crc-debug-pb5xg" Dec 09 11:16:04 crc kubenswrapper[4824]: I1209 11:16:04.721343 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/925ea289-5f58-4051-ab56-480312f713ef-host\") pod \"crc-debug-pb5xg\" (UID: \"925ea289-5f58-4051-ab56-480312f713ef\") " pod="openshift-must-gather-2w2mq/crc-debug-pb5xg" Dec 09 11:16:04 crc kubenswrapper[4824]: I1209 11:16:04.721513 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hkpn\" (UniqueName: \"kubernetes.io/projected/925ea289-5f58-4051-ab56-480312f713ef-kube-api-access-2hkpn\") pod \"crc-debug-pb5xg\" (UID: \"925ea289-5f58-4051-ab56-480312f713ef\") " pod="openshift-must-gather-2w2mq/crc-debug-pb5xg" Dec 09 11:16:04 crc kubenswrapper[4824]: I1209 11:16:04.722014 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/925ea289-5f58-4051-ab56-480312f713ef-host\") pod \"crc-debug-pb5xg\" (UID: \"925ea289-5f58-4051-ab56-480312f713ef\") " pod="openshift-must-gather-2w2mq/crc-debug-pb5xg" Dec 09 11:16:04 crc kubenswrapper[4824]: I1209 11:16:04.760699 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hkpn\" (UniqueName: \"kubernetes.io/projected/925ea289-5f58-4051-ab56-480312f713ef-kube-api-access-2hkpn\") pod \"crc-debug-pb5xg\" (UID: \"925ea289-5f58-4051-ab56-480312f713ef\") " pod="openshift-must-gather-2w2mq/crc-debug-pb5xg" Dec 09 11:16:04 crc kubenswrapper[4824]: I1209 11:16:04.907215 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2w2mq/crc-debug-pb5xg" Dec 09 11:16:05 crc kubenswrapper[4824]: I1209 11:16:05.913125 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:16:05 crc kubenswrapper[4824]: E1209 11:16:05.913848 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=ceilometer-central-agent pod=ceilometer-0_openstack(9339c10c-1cea-4ffc-8931-9b349a1e74cd)\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:16:05 crc kubenswrapper[4824]: I1209 11:16:05.963183 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2w2mq/crc-debug-pb5xg" event={"ID":"925ea289-5f58-4051-ab56-480312f713ef","Type":"ContainerStarted","Data":"f93932b7d3428f9daf2ced407f653e59a7f2061bb2415eb99f4c4ccc27b7b98f"} Dec 09 11:16:15 crc kubenswrapper[4824]: I1209 11:16:15.912315 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:16:15 crc kubenswrapper[4824]: E1209 11:16:15.913137 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:16:17 crc kubenswrapper[4824]: I1209 11:16:17.910846 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:16:17 crc kubenswrapper[4824]: E1209 11:16:17.911745 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=ceilometer-central-agent pod=ceilometer-0_openstack(9339c10c-1cea-4ffc-8931-9b349a1e74cd)\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:16:19 crc kubenswrapper[4824]: I1209 11:16:19.122210 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2w2mq/crc-debug-pb5xg" event={"ID":"925ea289-5f58-4051-ab56-480312f713ef","Type":"ContainerStarted","Data":"f74b8b8cf479040192843fd126417e43716736eae628c758f8682bb3d4cece17"} Dec 09 11:16:19 crc kubenswrapper[4824]: I1209 11:16:19.139151 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-2w2mq/crc-debug-pb5xg" podStartSLOduration=1.829270595 podStartE2EDuration="15.139133142s" podCreationTimestamp="2025-12-09 11:16:04 +0000 UTC" firstStartedPulling="2025-12-09 11:16:04.977198247 +0000 UTC m=+5321.311702914" lastFinishedPulling="2025-12-09 11:16:18.287060794 +0000 UTC m=+5334.621565461" observedRunningTime="2025-12-09 11:16:19.137650786 +0000 UTC m=+5335.472155463" watchObservedRunningTime="2025-12-09 11:16:19.139133142 +0000 UTC m=+5335.473637819" Dec 09 11:16:27 crc kubenswrapper[4824]: I1209 11:16:27.911174 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:16:28 crc kubenswrapper[4824]: E1209 11:16:27.912079 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:16:30 crc kubenswrapper[4824]: I1209 11:16:30.911692 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:16:30 crc kubenswrapper[4824]: E1209 11:16:30.912669 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=ceilometer-central-agent pod=ceilometer-0_openstack(9339c10c-1cea-4ffc-8931-9b349a1e74cd)\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:16:42 crc kubenswrapper[4824]: I1209 11:16:42.911200 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:16:42 crc kubenswrapper[4824]: E1209 11:16:42.912048 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:16:43 crc kubenswrapper[4824]: I1209 11:16:43.925540 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:16:43 crc kubenswrapper[4824]: E1209 11:16:43.926095 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=ceilometer-central-agent pod=ceilometer-0_openstack(9339c10c-1cea-4ffc-8931-9b349a1e74cd)\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:16:45 crc kubenswrapper[4824]: I1209 11:16:45.602528 4824 generic.go:334] "Generic (PLEG): container finished" podID="e153e30f-4f5e-4a38-8efb-6452096e25d7" containerID="7da26121a429a48f86bdeaadb323c671e73ba9993fa5bd40f3fcca78a5bf3769" exitCode=0 Dec 09 11:16:45 crc kubenswrapper[4824]: I1209 11:16:45.602587 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" event={"ID":"e153e30f-4f5e-4a38-8efb-6452096e25d7","Type":"ContainerDied","Data":"7da26121a429a48f86bdeaadb323c671e73ba9993fa5bd40f3fcca78a5bf3769"} Dec 09 11:16:45 crc kubenswrapper[4824]: I1209 11:16:45.603212 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" event={"ID":"e153e30f-4f5e-4a38-8efb-6452096e25d7","Type":"ContainerStarted","Data":"81a0497b98cd33fade04875390be2441f107f76ac4154682039ea6eecfb5713c"} Dec 09 11:16:55 crc kubenswrapper[4824]: I1209 11:16:55.911493 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:16:55 crc kubenswrapper[4824]: E1209 11:16:55.931595 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:16:57 crc kubenswrapper[4824]: I1209 11:16:57.916567 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:16:57 crc kubenswrapper[4824]: E1209 11:16:57.917192 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=ceilometer-central-agent pod=ceilometer-0_openstack(9339c10c-1cea-4ffc-8931-9b349a1e74cd)\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:17:01 crc kubenswrapper[4824]: I1209 11:17:01.513210 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 11:17:01 crc kubenswrapper[4824]: I1209 11:17:01.513981 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 11:17:07 crc kubenswrapper[4824]: I1209 11:17:07.142012 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:17:07 crc kubenswrapper[4824]: E1209 11:17:07.142867 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:17:10 crc kubenswrapper[4824]: I1209 11:17:10.911874 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:17:10 crc kubenswrapper[4824]: E1209 11:17:10.914230 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:17:17 crc kubenswrapper[4824]: I1209 11:17:17.910939 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:17:17 crc kubenswrapper[4824]: E1209 11:17:17.913010 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:17:18 crc kubenswrapper[4824]: I1209 11:17:18.290307 4824 generic.go:334] "Generic (PLEG): container finished" podID="925ea289-5f58-4051-ab56-480312f713ef" containerID="f74b8b8cf479040192843fd126417e43716736eae628c758f8682bb3d4cece17" exitCode=0 Dec 09 11:17:18 crc kubenswrapper[4824]: I1209 11:17:18.290357 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2w2mq/crc-debug-pb5xg" event={"ID":"925ea289-5f58-4051-ab56-480312f713ef","Type":"ContainerDied","Data":"f74b8b8cf479040192843fd126417e43716736eae628c758f8682bb3d4cece17"} Dec 09 11:17:19 crc kubenswrapper[4824]: I1209 11:17:19.449999 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2w2mq/crc-debug-pb5xg" Dec 09 11:17:19 crc kubenswrapper[4824]: I1209 11:17:19.499578 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2w2mq/crc-debug-pb5xg"] Dec 09 11:17:19 crc kubenswrapper[4824]: I1209 11:17:19.516513 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2w2mq/crc-debug-pb5xg"] Dec 09 11:17:19 crc kubenswrapper[4824]: I1209 11:17:19.619584 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/925ea289-5f58-4051-ab56-480312f713ef-host\") pod \"925ea289-5f58-4051-ab56-480312f713ef\" (UID: \"925ea289-5f58-4051-ab56-480312f713ef\") " Dec 09 11:17:19 crc kubenswrapper[4824]: I1209 11:17:19.620194 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/925ea289-5f58-4051-ab56-480312f713ef-host" (OuterVolumeSpecName: "host") pod "925ea289-5f58-4051-ab56-480312f713ef" (UID: "925ea289-5f58-4051-ab56-480312f713ef"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:17:19 crc kubenswrapper[4824]: I1209 11:17:19.620450 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hkpn\" (UniqueName: \"kubernetes.io/projected/925ea289-5f58-4051-ab56-480312f713ef-kube-api-access-2hkpn\") pod \"925ea289-5f58-4051-ab56-480312f713ef\" (UID: \"925ea289-5f58-4051-ab56-480312f713ef\") " Dec 09 11:17:19 crc kubenswrapper[4824]: I1209 11:17:19.621329 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/925ea289-5f58-4051-ab56-480312f713ef-host\") on node \"crc\" DevicePath \"\"" Dec 09 11:17:19 crc kubenswrapper[4824]: I1209 11:17:19.635220 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925ea289-5f58-4051-ab56-480312f713ef-kube-api-access-2hkpn" (OuterVolumeSpecName: "kube-api-access-2hkpn") pod "925ea289-5f58-4051-ab56-480312f713ef" (UID: "925ea289-5f58-4051-ab56-480312f713ef"). InnerVolumeSpecName "kube-api-access-2hkpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:17:19 crc kubenswrapper[4824]: I1209 11:17:19.723677 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hkpn\" (UniqueName: \"kubernetes.io/projected/925ea289-5f58-4051-ab56-480312f713ef-kube-api-access-2hkpn\") on node \"crc\" DevicePath \"\"" Dec 09 11:17:19 crc kubenswrapper[4824]: I1209 11:17:19.926372 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925ea289-5f58-4051-ab56-480312f713ef" path="/var/lib/kubelet/pods/925ea289-5f58-4051-ab56-480312f713ef/volumes" Dec 09 11:17:20 crc kubenswrapper[4824]: I1209 11:17:20.312810 4824 scope.go:117] "RemoveContainer" containerID="f74b8b8cf479040192843fd126417e43716736eae628c758f8682bb3d4cece17" Dec 09 11:17:20 crc kubenswrapper[4824]: I1209 11:17:20.312835 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2w2mq/crc-debug-pb5xg" Dec 09 11:17:20 crc kubenswrapper[4824]: I1209 11:17:20.735516 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2w2mq/crc-debug-f9xbv"] Dec 09 11:17:20 crc kubenswrapper[4824]: E1209 11:17:20.736304 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="925ea289-5f58-4051-ab56-480312f713ef" containerName="container-00" Dec 09 11:17:20 crc kubenswrapper[4824]: I1209 11:17:20.736329 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="925ea289-5f58-4051-ab56-480312f713ef" containerName="container-00" Dec 09 11:17:20 crc kubenswrapper[4824]: I1209 11:17:20.736567 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="925ea289-5f58-4051-ab56-480312f713ef" containerName="container-00" Dec 09 11:17:20 crc kubenswrapper[4824]: I1209 11:17:20.737598 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2w2mq/crc-debug-f9xbv" Dec 09 11:17:20 crc kubenswrapper[4824]: I1209 11:17:20.763440 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwnc9\" (UniqueName: \"kubernetes.io/projected/a979c686-43ed-4947-918b-e2390b12a8c3-kube-api-access-mwnc9\") pod \"crc-debug-f9xbv\" (UID: \"a979c686-43ed-4947-918b-e2390b12a8c3\") " pod="openshift-must-gather-2w2mq/crc-debug-f9xbv" Dec 09 11:17:20 crc kubenswrapper[4824]: I1209 11:17:20.763726 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a979c686-43ed-4947-918b-e2390b12a8c3-host\") pod \"crc-debug-f9xbv\" (UID: \"a979c686-43ed-4947-918b-e2390b12a8c3\") " pod="openshift-must-gather-2w2mq/crc-debug-f9xbv" Dec 09 11:17:20 crc kubenswrapper[4824]: I1209 11:17:20.865727 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a979c686-43ed-4947-918b-e2390b12a8c3-host\") pod \"crc-debug-f9xbv\" (UID: \"a979c686-43ed-4947-918b-e2390b12a8c3\") " pod="openshift-must-gather-2w2mq/crc-debug-f9xbv" Dec 09 11:17:20 crc kubenswrapper[4824]: I1209 11:17:20.865878 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a979c686-43ed-4947-918b-e2390b12a8c3-host\") pod \"crc-debug-f9xbv\" (UID: \"a979c686-43ed-4947-918b-e2390b12a8c3\") " pod="openshift-must-gather-2w2mq/crc-debug-f9xbv" Dec 09 11:17:20 crc kubenswrapper[4824]: I1209 11:17:20.865906 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwnc9\" (UniqueName: \"kubernetes.io/projected/a979c686-43ed-4947-918b-e2390b12a8c3-kube-api-access-mwnc9\") pod \"crc-debug-f9xbv\" (UID: \"a979c686-43ed-4947-918b-e2390b12a8c3\") " pod="openshift-must-gather-2w2mq/crc-debug-f9xbv" Dec 09 11:17:21 crc kubenswrapper[4824]: I1209 11:17:21.143842 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwnc9\" (UniqueName: \"kubernetes.io/projected/a979c686-43ed-4947-918b-e2390b12a8c3-kube-api-access-mwnc9\") pod \"crc-debug-f9xbv\" (UID: \"a979c686-43ed-4947-918b-e2390b12a8c3\") " pod="openshift-must-gather-2w2mq/crc-debug-f9xbv" Dec 09 11:17:21 crc kubenswrapper[4824]: I1209 11:17:21.358397 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2w2mq/crc-debug-f9xbv" Dec 09 11:17:21 crc kubenswrapper[4824]: I1209 11:17:21.518757 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 11:17:21 crc kubenswrapper[4824]: I1209 11:17:21.528690 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/metrics-server-65846dd9c4-g8tlp" Dec 09 11:17:22 crc kubenswrapper[4824]: I1209 11:17:22.335046 4824 generic.go:334] "Generic (PLEG): container finished" podID="a979c686-43ed-4947-918b-e2390b12a8c3" containerID="92646645dfb5954585aeb097e5cc57379442f6ddc90f89ca430e7013126fa39f" exitCode=0 Dec 09 11:17:22 crc kubenswrapper[4824]: I1209 11:17:22.335166 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2w2mq/crc-debug-f9xbv" event={"ID":"a979c686-43ed-4947-918b-e2390b12a8c3","Type":"ContainerDied","Data":"92646645dfb5954585aeb097e5cc57379442f6ddc90f89ca430e7013126fa39f"} Dec 09 11:17:22 crc kubenswrapper[4824]: I1209 11:17:22.335394 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2w2mq/crc-debug-f9xbv" event={"ID":"a979c686-43ed-4947-918b-e2390b12a8c3","Type":"ContainerStarted","Data":"e9f08e5219c2cebe527d74f5d21a87ebd58a96089c6b8a8bb2a352011dcabe20"} Dec 09 11:17:22 crc kubenswrapper[4824]: I1209 11:17:22.913211 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:17:22 crc kubenswrapper[4824]: E1209 11:17:22.913468 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=ceilometer-central-agent pod=ceilometer-0_openstack(9339c10c-1cea-4ffc-8931-9b349a1e74cd)\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:17:23 crc kubenswrapper[4824]: I1209 11:17:23.383836 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2w2mq/crc-debug-f9xbv"] Dec 09 11:17:23 crc kubenswrapper[4824]: I1209 11:17:23.400242 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2w2mq/crc-debug-f9xbv"] Dec 09 11:17:23 crc kubenswrapper[4824]: I1209 11:17:23.500134 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2w2mq/crc-debug-f9xbv" Dec 09 11:17:23 crc kubenswrapper[4824]: I1209 11:17:23.559276 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwnc9\" (UniqueName: \"kubernetes.io/projected/a979c686-43ed-4947-918b-e2390b12a8c3-kube-api-access-mwnc9\") pod \"a979c686-43ed-4947-918b-e2390b12a8c3\" (UID: \"a979c686-43ed-4947-918b-e2390b12a8c3\") " Dec 09 11:17:23 crc kubenswrapper[4824]: I1209 11:17:23.559377 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a979c686-43ed-4947-918b-e2390b12a8c3-host\") pod \"a979c686-43ed-4947-918b-e2390b12a8c3\" (UID: \"a979c686-43ed-4947-918b-e2390b12a8c3\") " Dec 09 11:17:23 crc kubenswrapper[4824]: I1209 11:17:23.560250 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a979c686-43ed-4947-918b-e2390b12a8c3-host" (OuterVolumeSpecName: "host") pod "a979c686-43ed-4947-918b-e2390b12a8c3" (UID: "a979c686-43ed-4947-918b-e2390b12a8c3"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:17:23 crc kubenswrapper[4824]: I1209 11:17:23.578186 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a979c686-43ed-4947-918b-e2390b12a8c3-kube-api-access-mwnc9" (OuterVolumeSpecName: "kube-api-access-mwnc9") pod "a979c686-43ed-4947-918b-e2390b12a8c3" (UID: "a979c686-43ed-4947-918b-e2390b12a8c3"). InnerVolumeSpecName "kube-api-access-mwnc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:17:23 crc kubenswrapper[4824]: I1209 11:17:23.662888 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwnc9\" (UniqueName: \"kubernetes.io/projected/a979c686-43ed-4947-918b-e2390b12a8c3-kube-api-access-mwnc9\") on node \"crc\" DevicePath \"\"" Dec 09 11:17:23 crc kubenswrapper[4824]: I1209 11:17:23.662930 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a979c686-43ed-4947-918b-e2390b12a8c3-host\") on node \"crc\" DevicePath \"\"" Dec 09 11:17:23 crc kubenswrapper[4824]: I1209 11:17:23.930375 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a979c686-43ed-4947-918b-e2390b12a8c3" path="/var/lib/kubelet/pods/a979c686-43ed-4947-918b-e2390b12a8c3/volumes" Dec 09 11:17:24 crc kubenswrapper[4824]: I1209 11:17:24.371149 4824 scope.go:117] "RemoveContainer" containerID="92646645dfb5954585aeb097e5cc57379442f6ddc90f89ca430e7013126fa39f" Dec 09 11:17:24 crc kubenswrapper[4824]: I1209 11:17:24.371376 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2w2mq/crc-debug-f9xbv" Dec 09 11:17:24 crc kubenswrapper[4824]: I1209 11:17:24.619144 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2w2mq/crc-debug-z2zbs"] Dec 09 11:17:24 crc kubenswrapper[4824]: E1209 11:17:24.620952 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a979c686-43ed-4947-918b-e2390b12a8c3" containerName="container-00" Dec 09 11:17:24 crc kubenswrapper[4824]: I1209 11:17:24.620974 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a979c686-43ed-4947-918b-e2390b12a8c3" containerName="container-00" Dec 09 11:17:24 crc kubenswrapper[4824]: I1209 11:17:24.621680 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a979c686-43ed-4947-918b-e2390b12a8c3" containerName="container-00" Dec 09 11:17:24 crc kubenswrapper[4824]: I1209 11:17:24.624238 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2w2mq/crc-debug-z2zbs" Dec 09 11:17:24 crc kubenswrapper[4824]: I1209 11:17:24.706602 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1d71b52a-7fb2-4c35-89b2-31f237a2d0d1-host\") pod \"crc-debug-z2zbs\" (UID: \"1d71b52a-7fb2-4c35-89b2-31f237a2d0d1\") " pod="openshift-must-gather-2w2mq/crc-debug-z2zbs" Dec 09 11:17:24 crc kubenswrapper[4824]: I1209 11:17:24.706738 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqjzc\" (UniqueName: \"kubernetes.io/projected/1d71b52a-7fb2-4c35-89b2-31f237a2d0d1-kube-api-access-nqjzc\") pod \"crc-debug-z2zbs\" (UID: \"1d71b52a-7fb2-4c35-89b2-31f237a2d0d1\") " pod="openshift-must-gather-2w2mq/crc-debug-z2zbs" Dec 09 11:17:24 crc kubenswrapper[4824]: I1209 11:17:24.809899 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1d71b52a-7fb2-4c35-89b2-31f237a2d0d1-host\") pod \"crc-debug-z2zbs\" (UID: \"1d71b52a-7fb2-4c35-89b2-31f237a2d0d1\") " pod="openshift-must-gather-2w2mq/crc-debug-z2zbs" Dec 09 11:17:24 crc kubenswrapper[4824]: I1209 11:17:24.810063 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqjzc\" (UniqueName: \"kubernetes.io/projected/1d71b52a-7fb2-4c35-89b2-31f237a2d0d1-kube-api-access-nqjzc\") pod \"crc-debug-z2zbs\" (UID: \"1d71b52a-7fb2-4c35-89b2-31f237a2d0d1\") " pod="openshift-must-gather-2w2mq/crc-debug-z2zbs" Dec 09 11:17:24 crc kubenswrapper[4824]: I1209 11:17:24.810190 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1d71b52a-7fb2-4c35-89b2-31f237a2d0d1-host\") pod \"crc-debug-z2zbs\" (UID: \"1d71b52a-7fb2-4c35-89b2-31f237a2d0d1\") " pod="openshift-must-gather-2w2mq/crc-debug-z2zbs" Dec 09 11:17:24 crc kubenswrapper[4824]: I1209 11:17:24.829958 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqjzc\" (UniqueName: \"kubernetes.io/projected/1d71b52a-7fb2-4c35-89b2-31f237a2d0d1-kube-api-access-nqjzc\") pod \"crc-debug-z2zbs\" (UID: \"1d71b52a-7fb2-4c35-89b2-31f237a2d0d1\") " pod="openshift-must-gather-2w2mq/crc-debug-z2zbs" Dec 09 11:17:24 crc kubenswrapper[4824]: I1209 11:17:24.950447 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2w2mq/crc-debug-z2zbs" Dec 09 11:17:24 crc kubenswrapper[4824]: W1209 11:17:24.996378 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d71b52a_7fb2_4c35_89b2_31f237a2d0d1.slice/crio-2333525c62029f4f18ecd1eeaf76745a3a405c77eb4312bd58f9f6de3e6cf3c7 WatchSource:0}: Error finding container 2333525c62029f4f18ecd1eeaf76745a3a405c77eb4312bd58f9f6de3e6cf3c7: Status 404 returned error can't find the container with id 2333525c62029f4f18ecd1eeaf76745a3a405c77eb4312bd58f9f6de3e6cf3c7 Dec 09 11:17:25 crc kubenswrapper[4824]: I1209 11:17:25.383069 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2w2mq/crc-debug-z2zbs" event={"ID":"1d71b52a-7fb2-4c35-89b2-31f237a2d0d1","Type":"ContainerStarted","Data":"2333525c62029f4f18ecd1eeaf76745a3a405c77eb4312bd58f9f6de3e6cf3c7"} Dec 09 11:17:26 crc kubenswrapper[4824]: E1209 11:17:26.209707 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda979c686_43ed_4947_918b_e2390b12a8c3.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:17:26 crc kubenswrapper[4824]: I1209 11:17:26.397650 4824 generic.go:334] "Generic (PLEG): container finished" podID="1d71b52a-7fb2-4c35-89b2-31f237a2d0d1" containerID="647401afbfcfe4790947ba3c4b84041df187e8a3c75d93229a15d0a4b5dbc783" exitCode=0 Dec 09 11:17:26 crc kubenswrapper[4824]: I1209 11:17:26.397955 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2w2mq/crc-debug-z2zbs" event={"ID":"1d71b52a-7fb2-4c35-89b2-31f237a2d0d1","Type":"ContainerDied","Data":"647401afbfcfe4790947ba3c4b84041df187e8a3c75d93229a15d0a4b5dbc783"} Dec 09 11:17:27 crc kubenswrapper[4824]: I1209 11:17:27.465189 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2w2mq/crc-debug-z2zbs"] Dec 09 11:17:27 crc kubenswrapper[4824]: I1209 11:17:27.478756 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2w2mq/crc-debug-z2zbs"] Dec 09 11:17:28 crc kubenswrapper[4824]: I1209 11:17:28.563020 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2w2mq/crc-debug-z2zbs" Dec 09 11:17:28 crc kubenswrapper[4824]: I1209 11:17:28.711327 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqjzc\" (UniqueName: \"kubernetes.io/projected/1d71b52a-7fb2-4c35-89b2-31f237a2d0d1-kube-api-access-nqjzc\") pod \"1d71b52a-7fb2-4c35-89b2-31f237a2d0d1\" (UID: \"1d71b52a-7fb2-4c35-89b2-31f237a2d0d1\") " Dec 09 11:17:28 crc kubenswrapper[4824]: I1209 11:17:28.711569 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1d71b52a-7fb2-4c35-89b2-31f237a2d0d1-host\") pod \"1d71b52a-7fb2-4c35-89b2-31f237a2d0d1\" (UID: \"1d71b52a-7fb2-4c35-89b2-31f237a2d0d1\") " Dec 09 11:17:28 crc kubenswrapper[4824]: I1209 11:17:28.711971 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d71b52a-7fb2-4c35-89b2-31f237a2d0d1-host" (OuterVolumeSpecName: "host") pod "1d71b52a-7fb2-4c35-89b2-31f237a2d0d1" (UID: "1d71b52a-7fb2-4c35-89b2-31f237a2d0d1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 11:17:28 crc kubenswrapper[4824]: I1209 11:17:28.712166 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1d71b52a-7fb2-4c35-89b2-31f237a2d0d1-host\") on node \"crc\" DevicePath \"\"" Dec 09 11:17:29 crc kubenswrapper[4824]: I1209 11:17:29.145052 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d71b52a-7fb2-4c35-89b2-31f237a2d0d1-kube-api-access-nqjzc" (OuterVolumeSpecName: "kube-api-access-nqjzc") pod "1d71b52a-7fb2-4c35-89b2-31f237a2d0d1" (UID: "1d71b52a-7fb2-4c35-89b2-31f237a2d0d1"). InnerVolumeSpecName "kube-api-access-nqjzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:17:29 crc kubenswrapper[4824]: I1209 11:17:29.224324 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqjzc\" (UniqueName: \"kubernetes.io/projected/1d71b52a-7fb2-4c35-89b2-31f237a2d0d1-kube-api-access-nqjzc\") on node \"crc\" DevicePath \"\"" Dec 09 11:17:29 crc kubenswrapper[4824]: I1209 11:17:29.442365 4824 scope.go:117] "RemoveContainer" containerID="647401afbfcfe4790947ba3c4b84041df187e8a3c75d93229a15d0a4b5dbc783" Dec 09 11:17:29 crc kubenswrapper[4824]: I1209 11:17:29.442454 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2w2mq/crc-debug-z2zbs" Dec 09 11:17:29 crc kubenswrapper[4824]: I1209 11:17:29.911955 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:17:29 crc kubenswrapper[4824]: E1209 11:17:29.912394 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:17:29 crc kubenswrapper[4824]: I1209 11:17:29.927067 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d71b52a-7fb2-4c35-89b2-31f237a2d0d1" path="/var/lib/kubelet/pods/1d71b52a-7fb2-4c35-89b2-31f237a2d0d1/volumes" Dec 09 11:17:31 crc kubenswrapper[4824]: E1209 11:17:31.306199 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda979c686_43ed_4947_918b_e2390b12a8c3.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:17:33 crc kubenswrapper[4824]: I1209 11:17:33.937942 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:17:33 crc kubenswrapper[4824]: E1209 11:17:33.938856 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=ceilometer-central-agent pod=ceilometer-0_openstack(9339c10c-1cea-4ffc-8931-9b349a1e74cd)\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:17:36 crc kubenswrapper[4824]: E1209 11:17:36.528594 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda979c686_43ed_4947_918b_e2390b12a8c3.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:17:43 crc kubenswrapper[4824]: I1209 11:17:43.919625 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:17:43 crc kubenswrapper[4824]: E1209 11:17:43.920507 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:17:46 crc kubenswrapper[4824]: E1209 11:17:46.408403 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda979c686_43ed_4947_918b_e2390b12a8c3.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:17:46 crc kubenswrapper[4824]: E1209 11:17:46.590971 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda979c686_43ed_4947_918b_e2390b12a8c3.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:17:47 crc kubenswrapper[4824]: I1209 11:17:47.911381 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:17:47 crc kubenswrapper[4824]: E1209 11:17:47.912121 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=ceilometer-central-agent pod=ceilometer-0_openstack(9339c10c-1cea-4ffc-8931-9b349a1e74cd)\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:17:48 crc kubenswrapper[4824]: E1209 11:17:48.105813 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda979c686_43ed_4947_918b_e2390b12a8c3.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:17:48 crc kubenswrapper[4824]: E1209 11:17:48.106287 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda979c686_43ed_4947_918b_e2390b12a8c3.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:17:56 crc kubenswrapper[4824]: E1209 11:17:56.894901 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda979c686_43ed_4947_918b_e2390b12a8c3.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:17:57 crc kubenswrapper[4824]: I1209 11:17:57.669304 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_af7caa74-1882-4e75-9c52-942df2e40efe/aodh-api/0.log" Dec 09 11:17:57 crc kubenswrapper[4824]: I1209 11:17:57.878251 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_af7caa74-1882-4e75-9c52-942df2e40efe/aodh-evaluator/0.log" Dec 09 11:17:57 crc kubenswrapper[4824]: I1209 11:17:57.911392 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:17:57 crc kubenswrapper[4824]: E1209 11:17:57.911739 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:17:57 crc kubenswrapper[4824]: I1209 11:17:57.950079 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_af7caa74-1882-4e75-9c52-942df2e40efe/aodh-listener/0.log" Dec 09 11:17:57 crc kubenswrapper[4824]: I1209 11:17:57.953242 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_af7caa74-1882-4e75-9c52-942df2e40efe/aodh-notifier/0.log" Dec 09 11:17:58 crc kubenswrapper[4824]: I1209 11:17:58.106075 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-867bb6d64d-9slcv_4dc240ff-1bf1-464e-8733-59d0ae2fc4d2/barbican-api/0.log" Dec 09 11:17:58 crc kubenswrapper[4824]: I1209 11:17:58.184118 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-867bb6d64d-9slcv_4dc240ff-1bf1-464e-8733-59d0ae2fc4d2/barbican-api-log/0.log" Dec 09 11:17:58 crc kubenswrapper[4824]: I1209 11:17:58.407748 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-55b96fff66-brzz9_d72ff513-d8d9-4191-92d4-7824fca18f32/barbican-keystone-listener/0.log" Dec 09 11:17:58 crc kubenswrapper[4824]: I1209 11:17:58.487688 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-55b96fff66-brzz9_d72ff513-d8d9-4191-92d4-7824fca18f32/barbican-keystone-listener-log/0.log" Dec 09 11:17:58 crc kubenswrapper[4824]: I1209 11:17:58.591864 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-77bcf58d55-bshs5_555ef974-afb9-4455-8a15-c728200be8a2/barbican-worker/0.log" Dec 09 11:17:58 crc kubenswrapper[4824]: I1209 11:17:58.675093 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-77bcf58d55-bshs5_555ef974-afb9-4455-8a15-c728200be8a2/barbican-worker-log/0.log" Dec 09 11:17:58 crc kubenswrapper[4824]: I1209 11:17:58.803023 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-q47fq_88cab327-4ed1-492c-9fd1-11a1657ec22f/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 11:17:58 crc kubenswrapper[4824]: I1209 11:17:58.982532 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9339c10c-1cea-4ffc-8931-9b349a1e74cd/ceilometer-central-agent/0.log" Dec 09 11:17:59 crc kubenswrapper[4824]: I1209 11:17:59.083151 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9339c10c-1cea-4ffc-8931-9b349a1e74cd/ceilometer-notification-agent/0.log" Dec 09 11:17:59 crc kubenswrapper[4824]: I1209 11:17:59.097713 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9339c10c-1cea-4ffc-8931-9b349a1e74cd/ceilometer-central-agent/0.log" Dec 09 11:17:59 crc kubenswrapper[4824]: I1209 11:17:59.134136 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9339c10c-1cea-4ffc-8931-9b349a1e74cd/proxy-httpd/0.log" Dec 09 11:17:59 crc kubenswrapper[4824]: I1209 11:17:59.220341 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9339c10c-1cea-4ffc-8931-9b349a1e74cd/sg-core/0.log" Dec 09 11:17:59 crc kubenswrapper[4824]: I1209 11:17:59.355175 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e0755044-204f-4e63-a3cb-69b7c8b64159/cinder-api/0.log" Dec 09 11:17:59 crc kubenswrapper[4824]: I1209 11:17:59.410300 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e0755044-204f-4e63-a3cb-69b7c8b64159/cinder-api-log/0.log" Dec 09 11:17:59 crc kubenswrapper[4824]: I1209 11:17:59.614280 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_153fe73c-2767-4f80-a7b4-0af49257065a/cinder-scheduler/1.log" Dec 09 11:17:59 crc kubenswrapper[4824]: I1209 11:17:59.626726 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_153fe73c-2767-4f80-a7b4-0af49257065a/cinder-scheduler/0.log" Dec 09 11:17:59 crc kubenswrapper[4824]: I1209 11:17:59.701831 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_153fe73c-2767-4f80-a7b4-0af49257065a/probe/0.log" Dec 09 11:17:59 crc kubenswrapper[4824]: I1209 11:17:59.859205 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-6z8pq_329f5680-bbd1-427c-b365-70c2109c214f/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 11:17:59 crc kubenswrapper[4824]: I1209 11:17:59.969254 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-fr8bs_721d01dd-ea5d-41e9-a01c-c6fe4ba28086/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 11:18:00 crc kubenswrapper[4824]: I1209 11:18:00.062973 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-2vs9q_2510a71c-bf12-4207-96ee-6722c9d2febe/init/0.log" Dec 09 11:18:00 crc kubenswrapper[4824]: I1209 11:18:00.319067 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-2vs9q_2510a71c-bf12-4207-96ee-6722c9d2febe/init/0.log" Dec 09 11:18:00 crc kubenswrapper[4824]: I1209 11:18:00.873146 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_22dc07da-136d-4812-b6c4-b7fa94903b5e/glance-log/0.log" Dec 09 11:18:00 crc kubenswrapper[4824]: I1209 11:18:00.926511 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-l7wgs_4a2f92dc-7710-4e22-8e1a-d303580314c2/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 11:18:00 crc kubenswrapper[4824]: I1209 11:18:00.936287 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_22dc07da-136d-4812-b6c4-b7fa94903b5e/glance-httpd/0.log" Dec 09 11:18:00 crc kubenswrapper[4824]: I1209 11:18:00.957894 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-2vs9q_2510a71c-bf12-4207-96ee-6722c9d2febe/dnsmasq-dns/0.log" Dec 09 11:18:01 crc kubenswrapper[4824]: E1209 11:18:01.073414 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda979c686_43ed_4947_918b_e2390b12a8c3.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:18:01 crc kubenswrapper[4824]: I1209 11:18:01.203667 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_5a1be792-ec64-4666-ada9-1ca72712213f/glance-log/0.log" Dec 09 11:18:01 crc kubenswrapper[4824]: I1209 11:18:01.255100 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_5a1be792-ec64-4666-ada9-1ca72712213f/glance-httpd/0.log" Dec 09 11:18:02 crc kubenswrapper[4824]: I1209 11:18:02.164016 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-67ffc9d677-6k2l9_ce017433-0624-4b3f-b1e9-0de3b591a602/heat-engine/0.log" Dec 09 11:18:02 crc kubenswrapper[4824]: I1209 11:18:02.294078 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-69c99fc8cd-j8cz5_1ec002ca-0e91-48ab-a982-51d6c19b33b6/heat-cfnapi/0.log" Dec 09 11:18:02 crc kubenswrapper[4824]: I1209 11:18:02.312678 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-lncts_332799e5-f463-48ab-be97-77329e48b07d/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 11:18:02 crc kubenswrapper[4824]: I1209 11:18:02.326952 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-57cb8dfdcd-c62t6_aee6f3d2-13da-489e-844c-70355d34d102/heat-api/0.log" Dec 09 11:18:02 crc kubenswrapper[4824]: I1209 11:18:02.567661 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-7m2lk_a872b5e1-5510-4c90-bd4b-ebaf84dae414/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 11:18:02 crc kubenswrapper[4824]: I1209 11:18:02.770669 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29421301-s69hp_818d4716-9337-4843-ac3b-9247722f5ef0/keystone-cron/0.log" Dec 09 11:18:02 crc kubenswrapper[4824]: I1209 11:18:02.910341 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:18:02 crc kubenswrapper[4824]: E1209 11:18:02.910655 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=ceilometer-central-agent pod=ceilometer-0_openstack(9339c10c-1cea-4ffc-8931-9b349a1e74cd)\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:18:02 crc kubenswrapper[4824]: I1209 11:18:02.919645 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_54621281-db00-41e4-b617-032435893391/kube-state-metrics/1.log" Dec 09 11:18:02 crc kubenswrapper[4824]: I1209 11:18:02.951600 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_54621281-db00-41e4-b617-032435893391/kube-state-metrics/0.log" Dec 09 11:18:03 crc kubenswrapper[4824]: I1209 11:18:03.231150 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_logging-edpm-deployment-openstack-edpm-ipam-hj749_edcacc59-d95a-42e3-8bc4-9e8a1a7df2be/logging-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 11:18:03 crc kubenswrapper[4824]: I1209 11:18:03.298902 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-7jcnk_f2df5cd6-d375-4545-bf5a-bd3df55fb736/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 11:18:03 crc kubenswrapper[4824]: I1209 11:18:03.667827 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mysqld-exporter-0_674c8cb5-4f45-48a9-9b64-07ce6e0a0865/mysqld-exporter/0.log" Dec 09 11:18:04 crc kubenswrapper[4824]: I1209 11:18:04.165285 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7d56cf488f-6wtnd_ef937030-1ed4-4462-b362-aa335e77a9f6/neutron-httpd/0.log" Dec 09 11:18:04 crc kubenswrapper[4824]: I1209 11:18:04.168925 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7d56cf488f-6wtnd_ef937030-1ed4-4462-b362-aa335e77a9f6/neutron-api/0.log" Dec 09 11:18:04 crc kubenswrapper[4824]: I1209 11:18:04.212976 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-9c7879457-2fmzw_9dcb334f-65bf-4f84-88a0-24ca56b4a8c7/keystone-api/0.log" Dec 09 11:18:04 crc kubenswrapper[4824]: I1209 11:18:04.408359 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-c9t8x_3fd9b8ef-292c-46ee-938e-62daa3bda9e0/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 11:18:04 crc kubenswrapper[4824]: I1209 11:18:04.953994 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_9fe3dc96-b1f9-475b-9ad7-e752f0cc434e/nova-api-log/0.log" Dec 09 11:18:05 crc kubenswrapper[4824]: I1209 11:18:05.065174 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_71b5e2f1-303e-4718-be1f-fb2ab839dd64/nova-cell0-conductor-conductor/0.log" Dec 09 11:18:05 crc kubenswrapper[4824]: I1209 11:18:05.349940 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_a2728556-b7df-4f3b-b3d7-ec6f2683b0fa/nova-cell1-conductor-conductor/0.log" Dec 09 11:18:05 crc kubenswrapper[4824]: I1209 11:18:05.398499 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_9fe3dc96-b1f9-475b-9ad7-e752f0cc434e/nova-api-api/0.log" Dec 09 11:18:05 crc kubenswrapper[4824]: I1209 11:18:05.478946 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_a6208994-9aba-4007-ba54-b8718e789f4b/nova-cell1-novncproxy-novncproxy/0.log" Dec 09 11:18:05 crc kubenswrapper[4824]: I1209 11:18:05.655093 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-7v7ms_facb5571-0efc-4b89-8f5a-69dcff002fa5/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 11:18:05 crc kubenswrapper[4824]: I1209 11:18:05.804372 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_b295ca7f-46ac-4ab1-afc4-f1554b84c0d3/nova-metadata-log/0.log" Dec 09 11:18:06 crc kubenswrapper[4824]: I1209 11:18:06.206588 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_543b3a99-2d5c-4467-8076-b284ea519244/nova-scheduler-scheduler/0.log" Dec 09 11:18:06 crc kubenswrapper[4824]: I1209 11:18:06.213146 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_73d6bd70-44c7-4eed-a93a-36df636869cf/mysql-bootstrap/0.log" Dec 09 11:18:06 crc kubenswrapper[4824]: I1209 11:18:06.483764 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_73d6bd70-44c7-4eed-a93a-36df636869cf/galera/1.log" Dec 09 11:18:06 crc kubenswrapper[4824]: I1209 11:18:06.493299 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_73d6bd70-44c7-4eed-a93a-36df636869cf/mysql-bootstrap/0.log" Dec 09 11:18:06 crc kubenswrapper[4824]: I1209 11:18:06.519224 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_73d6bd70-44c7-4eed-a93a-36df636869cf/galera/0.log" Dec 09 11:18:06 crc kubenswrapper[4824]: I1209 11:18:06.741723 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_56148e0f-636f-410e-bfb5-342da01e8c76/mysql-bootstrap/0.log" Dec 09 11:18:06 crc kubenswrapper[4824]: I1209 11:18:06.923582 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_56148e0f-636f-410e-bfb5-342da01e8c76/mysql-bootstrap/0.log" Dec 09 11:18:06 crc kubenswrapper[4824]: E1209 11:18:06.948603 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda979c686_43ed_4947_918b_e2390b12a8c3.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:18:06 crc kubenswrapper[4824]: I1209 11:18:06.995654 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_56148e0f-636f-410e-bfb5-342da01e8c76/galera/0.log" Dec 09 11:18:07 crc kubenswrapper[4824]: I1209 11:18:07.017364 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_56148e0f-636f-410e-bfb5-342da01e8c76/galera/1.log" Dec 09 11:18:07 crc kubenswrapper[4824]: I1209 11:18:07.214280 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_a00596c6-582b-4ec7-9920-6dddc8197001/openstackclient/0.log" Dec 09 11:18:07 crc kubenswrapper[4824]: I1209 11:18:07.556407 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-985bt_961b1a8a-2fa7-4b53-bf91-c28a6d2fea6d/ovn-controller/0.log" Dec 09 11:18:07 crc kubenswrapper[4824]: I1209 11:18:07.778118 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-6f874_c0ec6983-7e9d-449f-99ce-44b0bc708f19/openstack-network-exporter/0.log" Dec 09 11:18:07 crc kubenswrapper[4824]: I1209 11:18:07.893408 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_b295ca7f-46ac-4ab1-afc4-f1554b84c0d3/nova-metadata-metadata/0.log" Dec 09 11:18:07 crc kubenswrapper[4824]: I1209 11:18:07.916604 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rl4nw_2a54d12b-327d-409e-8652-9525878ae96f/ovsdb-server-init/0.log" Dec 09 11:18:08 crc kubenswrapper[4824]: I1209 11:18:08.117495 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rl4nw_2a54d12b-327d-409e-8652-9525878ae96f/ovsdb-server-init/0.log" Dec 09 11:18:08 crc kubenswrapper[4824]: I1209 11:18:08.126282 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rl4nw_2a54d12b-327d-409e-8652-9525878ae96f/ovs-vswitchd/0.log" Dec 09 11:18:08 crc kubenswrapper[4824]: I1209 11:18:08.159429 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rl4nw_2a54d12b-327d-409e-8652-9525878ae96f/ovsdb-server/0.log" Dec 09 11:18:08 crc kubenswrapper[4824]: I1209 11:18:08.178227 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_4afe9ced-d6b6-4684-96f3-a9a419711cc7/memcached/0.log" Dec 09 11:18:08 crc kubenswrapper[4824]: I1209 11:18:08.346386 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c2dd0680-d32f-45e8-b834-6a327d2eaa21/openstack-network-exporter/0.log" Dec 09 11:18:08 crc kubenswrapper[4824]: I1209 11:18:08.365238 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c2dd0680-d32f-45e8-b834-6a327d2eaa21/ovn-northd/1.log" Dec 09 11:18:08 crc kubenswrapper[4824]: I1209 11:18:08.365417 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-dvksh_5c209e08-d978-483b-88b8-2f0ef683d62e/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 11:18:08 crc kubenswrapper[4824]: I1209 11:18:08.432228 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c2dd0680-d32f-45e8-b834-6a327d2eaa21/ovn-northd/0.log" Dec 09 11:18:08 crc kubenswrapper[4824]: I1209 11:18:08.558985 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c22f05c5-6c0e-4e80-afff-b2642d4a7655/openstack-network-exporter/0.log" Dec 09 11:18:08 crc kubenswrapper[4824]: I1209 11:18:08.566031 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c22f05c5-6c0e-4e80-afff-b2642d4a7655/ovsdbserver-nb/0.log" Dec 09 11:18:08 crc kubenswrapper[4824]: I1209 11:18:08.658039 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ec252d15-4814-478d-ad29-388842d50a32/openstack-network-exporter/0.log" Dec 09 11:18:08 crc kubenswrapper[4824]: I1209 11:18:08.737471 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ec252d15-4814-478d-ad29-388842d50a32/ovsdbserver-sb/0.log" Dec 09 11:18:08 crc kubenswrapper[4824]: I1209 11:18:08.899701 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7c7958c67b-wr8zx_38c32e83-50e7-4f8b-8bbf-26adebff1b82/placement-api/0.log" Dec 09 11:18:08 crc kubenswrapper[4824]: I1209 11:18:08.989797 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_fb8f1e44-4c21-423a-bacc-1cca7d7715c3/init-config-reloader/0.log" Dec 09 11:18:09 crc kubenswrapper[4824]: I1209 11:18:09.016175 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7c7958c67b-wr8zx_38c32e83-50e7-4f8b-8bbf-26adebff1b82/placement-log/0.log" Dec 09 11:18:09 crc kubenswrapper[4824]: I1209 11:18:09.190215 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_fb8f1e44-4c21-423a-bacc-1cca7d7715c3/prometheus/0.log" Dec 09 11:18:09 crc kubenswrapper[4824]: I1209 11:18:09.212552 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_fb8f1e44-4c21-423a-bacc-1cca7d7715c3/config-reloader/0.log" Dec 09 11:18:09 crc kubenswrapper[4824]: I1209 11:18:09.217607 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_fb8f1e44-4c21-423a-bacc-1cca7d7715c3/init-config-reloader/0.log" Dec 09 11:18:09 crc kubenswrapper[4824]: I1209 11:18:09.291685 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_fb8f1e44-4c21-423a-bacc-1cca7d7715c3/thanos-sidecar/0.log" Dec 09 11:18:09 crc kubenswrapper[4824]: I1209 11:18:09.375618 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b3de6262-0cdb-411a-9da1-9815c40d577e/setup-container/0.log" Dec 09 11:18:09 crc kubenswrapper[4824]: I1209 11:18:09.590896 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b3de6262-0cdb-411a-9da1-9815c40d577e/setup-container/0.log" Dec 09 11:18:09 crc kubenswrapper[4824]: I1209 11:18:09.604416 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b3de6262-0cdb-411a-9da1-9815c40d577e/rabbitmq/0.log" Dec 09 11:18:09 crc kubenswrapper[4824]: I1209 11:18:09.633886 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_42e63d7a-a5a7-47e3-933f-8e868262c6d6/setup-container/0.log" Dec 09 11:18:09 crc kubenswrapper[4824]: I1209 11:18:09.825776 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_42e63d7a-a5a7-47e3-933f-8e868262c6d6/setup-container/0.log" Dec 09 11:18:09 crc kubenswrapper[4824]: I1209 11:18:09.859748 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-1_5de00a63-1140-413d-be79-922acdae88bc/setup-container/0.log" Dec 09 11:18:09 crc kubenswrapper[4824]: I1209 11:18:09.872324 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_42e63d7a-a5a7-47e3-933f-8e868262c6d6/rabbitmq/0.log" Dec 09 11:18:10 crc kubenswrapper[4824]: I1209 11:18:10.080859 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-1_5de00a63-1140-413d-be79-922acdae88bc/setup-container/0.log" Dec 09 11:18:10 crc kubenswrapper[4824]: I1209 11:18:10.191591 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-1_5de00a63-1140-413d-be79-922acdae88bc/rabbitmq/0.log" Dec 09 11:18:10 crc kubenswrapper[4824]: I1209 11:18:10.200581 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-2_c97a0c28-3a5a-430a-8000-e5c0518262b3/setup-container/0.log" Dec 09 11:18:10 crc kubenswrapper[4824]: I1209 11:18:10.355975 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-2_c97a0c28-3a5a-430a-8000-e5c0518262b3/rabbitmq/0.log" Dec 09 11:18:10 crc kubenswrapper[4824]: I1209 11:18:10.364988 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-2_c97a0c28-3a5a-430a-8000-e5c0518262b3/setup-container/0.log" Dec 09 11:18:10 crc kubenswrapper[4824]: I1209 11:18:10.451730 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-w7rt6_21d8d8f6-2356-406f-a811-e867f326d758/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 11:18:10 crc kubenswrapper[4824]: I1209 11:18:10.587617 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-vbfns_eeb8228d-77ed-45ad-b2dc-085a2ba8814e/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 11:18:10 crc kubenswrapper[4824]: I1209 11:18:10.732917 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-nqb2j_6e1db872-7d04-4c89-b725-9e3d84fed0fb/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 11:18:10 crc kubenswrapper[4824]: I1209 11:18:10.751112 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-d5rkl_b3329105-6b83-49ff-8dd7-5343a656f7af/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 11:18:10 crc kubenswrapper[4824]: I1209 11:18:10.910681 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:18:10 crc kubenswrapper[4824]: E1209 11:18:10.910986 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:18:11 crc kubenswrapper[4824]: I1209 11:18:11.779121 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-85dcdb56bc-4xfwk_574c42ce-f75f-4dab-bd77-916b9a4d2c6d/proxy-server/0.log" Dec 09 11:18:11 crc kubenswrapper[4824]: I1209 11:18:11.789419 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-6tlwx_eae91833-e890-428b-beee-3420b3718f45/ssh-known-hosts-edpm-deployment/0.log" Dec 09 11:18:11 crc kubenswrapper[4824]: I1209 11:18:11.848618 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-85dcdb56bc-4xfwk_574c42ce-f75f-4dab-bd77-916b9a4d2c6d/proxy-httpd/0.log" Dec 09 11:18:11 crc kubenswrapper[4824]: I1209 11:18:11.948542 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-cf5wn_abbc90b3-afb8-4c42-a07d-1d1c38b8d87c/swift-ring-rebalance/0.log" Dec 09 11:18:12 crc kubenswrapper[4824]: I1209 11:18:12.064137 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66fd04f1-a0d9-4295-ac42-e5b20acc7c1f/account-auditor/0.log" Dec 09 11:18:12 crc kubenswrapper[4824]: I1209 11:18:12.102180 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66fd04f1-a0d9-4295-ac42-e5b20acc7c1f/account-reaper/0.log" Dec 09 11:18:12 crc kubenswrapper[4824]: I1209 11:18:12.194674 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66fd04f1-a0d9-4295-ac42-e5b20acc7c1f/account-replicator/0.log" Dec 09 11:18:12 crc kubenswrapper[4824]: I1209 11:18:12.253184 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66fd04f1-a0d9-4295-ac42-e5b20acc7c1f/account-server/0.log" Dec 09 11:18:12 crc kubenswrapper[4824]: I1209 11:18:12.300878 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66fd04f1-a0d9-4295-ac42-e5b20acc7c1f/container-auditor/0.log" Dec 09 11:18:12 crc kubenswrapper[4824]: I1209 11:18:12.313417 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66fd04f1-a0d9-4295-ac42-e5b20acc7c1f/container-server/0.log" Dec 09 11:18:12 crc kubenswrapper[4824]: I1209 11:18:12.338050 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66fd04f1-a0d9-4295-ac42-e5b20acc7c1f/container-replicator/0.log" Dec 09 11:18:12 crc kubenswrapper[4824]: I1209 11:18:12.428543 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66fd04f1-a0d9-4295-ac42-e5b20acc7c1f/container-updater/0.log" Dec 09 11:18:12 crc kubenswrapper[4824]: I1209 11:18:12.469170 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66fd04f1-a0d9-4295-ac42-e5b20acc7c1f/object-auditor/0.log" Dec 09 11:18:12 crc kubenswrapper[4824]: I1209 11:18:12.524043 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66fd04f1-a0d9-4295-ac42-e5b20acc7c1f/object-expirer/0.log" Dec 09 11:18:12 crc kubenswrapper[4824]: I1209 11:18:12.567749 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66fd04f1-a0d9-4295-ac42-e5b20acc7c1f/object-replicator/0.log" Dec 09 11:18:12 crc kubenswrapper[4824]: I1209 11:18:12.573220 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66fd04f1-a0d9-4295-ac42-e5b20acc7c1f/object-server/0.log" Dec 09 11:18:12 crc kubenswrapper[4824]: I1209 11:18:12.661305 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66fd04f1-a0d9-4295-ac42-e5b20acc7c1f/object-updater/0.log" Dec 09 11:18:12 crc kubenswrapper[4824]: I1209 11:18:12.719273 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66fd04f1-a0d9-4295-ac42-e5b20acc7c1f/rsync/0.log" Dec 09 11:18:12 crc kubenswrapper[4824]: I1209 11:18:12.757279 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_66fd04f1-a0d9-4295-ac42-e5b20acc7c1f/swift-recon-cron/0.log" Dec 09 11:18:12 crc kubenswrapper[4824]: I1209 11:18:12.826220 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-grqfl_86bfd02c-d226-48e8-bace-e961440bdb43/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 11:18:14 crc kubenswrapper[4824]: I1209 11:18:14.054000 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-power-monitoring-edpm-deployment-openstack-edpm-4bm26_c62caaf5-b010-47ee-9987-50cd9b418a28/telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 11:18:14 crc kubenswrapper[4824]: I1209 11:18:14.200926 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_d29986ea-5ca9-442e-999b-48ac99ddcc9e/test-operator-logs-container/0.log" Dec 09 11:18:14 crc kubenswrapper[4824]: I1209 11:18:14.467132 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-stmd9_00da1369-ca64-4df7-9924-fbcd75b38d86/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 11:18:14 crc kubenswrapper[4824]: I1209 11:18:14.594030 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_0fd5955c-4039-42a1-8ecd-0256bb683e61/tempest-tests-tempest-tests-runner/0.log" Dec 09 11:18:16 crc kubenswrapper[4824]: E1209 11:18:16.390537 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda979c686_43ed_4947_918b_e2390b12a8c3.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:18:17 crc kubenswrapper[4824]: E1209 11:18:17.004585 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda979c686_43ed_4947_918b_e2390b12a8c3.slice\": RecentStats: unable to find data in memory cache]" Dec 09 11:18:17 crc kubenswrapper[4824]: I1209 11:18:17.911387 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:18:17 crc kubenswrapper[4824]: E1209 11:18:17.912975 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=ceilometer-central-agent pod=ceilometer-0_openstack(9339c10c-1cea-4ffc-8931-9b349a1e74cd)\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:18:25 crc kubenswrapper[4824]: I1209 11:18:25.912237 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:18:25 crc kubenswrapper[4824]: E1209 11:18:25.913158 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:18:30 crc kubenswrapper[4824]: I1209 11:18:30.910854 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:18:30 crc kubenswrapper[4824]: E1209 11:18:30.911842 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=ceilometer-central-agent pod=ceilometer-0_openstack(9339c10c-1cea-4ffc-8931-9b349a1e74cd)\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:18:40 crc kubenswrapper[4824]: I1209 11:18:40.911301 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:18:40 crc kubenswrapper[4824]: E1209 11:18:40.912267 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:18:42 crc kubenswrapper[4824]: I1209 11:18:42.467092 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-6mfg8_5b34da01-28d8-476c-a2a5-a098489d90c1/manager/0.log" Dec 09 11:18:42 crc kubenswrapper[4824]: I1209 11:18:42.474807 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-6mfg8_5b34da01-28d8-476c-a2a5-a098489d90c1/kube-rbac-proxy/0.log" Dec 09 11:18:42 crc kubenswrapper[4824]: I1209 11:18:42.690368 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-lpf4s_270a1cf5-f943-4749-8c9a-1c6adf3c45b4/kube-rbac-proxy/0.log" Dec 09 11:18:42 crc kubenswrapper[4824]: I1209 11:18:42.742879 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-lpf4s_270a1cf5-f943-4749-8c9a-1c6adf3c45b4/manager/0.log" Dec 09 11:18:42 crc kubenswrapper[4824]: I1209 11:18:42.805362 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk_6910cb67-1871-410a-bc4e-91289bed7d5a/util/0.log" Dec 09 11:18:43 crc kubenswrapper[4824]: I1209 11:18:43.018426 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk_6910cb67-1871-410a-bc4e-91289bed7d5a/pull/0.log" Dec 09 11:18:43 crc kubenswrapper[4824]: I1209 11:18:43.021154 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk_6910cb67-1871-410a-bc4e-91289bed7d5a/util/0.log" Dec 09 11:18:43 crc kubenswrapper[4824]: I1209 11:18:43.021711 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk_6910cb67-1871-410a-bc4e-91289bed7d5a/pull/0.log" Dec 09 11:18:43 crc kubenswrapper[4824]: I1209 11:18:43.199125 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk_6910cb67-1871-410a-bc4e-91289bed7d5a/pull/0.log" Dec 09 11:18:43 crc kubenswrapper[4824]: I1209 11:18:43.215419 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk_6910cb67-1871-410a-bc4e-91289bed7d5a/util/0.log" Dec 09 11:18:43 crc kubenswrapper[4824]: I1209 11:18:43.245536 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d5396dc8446af86c38e0755d235b4cf611925ba4a40ee3a4d423a614d84g2rk_6910cb67-1871-410a-bc4e-91289bed7d5a/extract/0.log" Dec 09 11:18:43 crc kubenswrapper[4824]: I1209 11:18:43.415345 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-n45n4_143e48cd-c956-448c-8dcb-5858a582e29c/manager/0.log" Dec 09 11:18:43 crc kubenswrapper[4824]: I1209 11:18:43.441100 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-n45n4_143e48cd-c956-448c-8dcb-5858a582e29c/manager/1.log" Dec 09 11:18:43 crc kubenswrapper[4824]: I1209 11:18:43.449941 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-n45n4_143e48cd-c956-448c-8dcb-5858a582e29c/kube-rbac-proxy/0.log" Dec 09 11:18:43 crc kubenswrapper[4824]: I1209 11:18:43.600121 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-vx285_54693584-08f2-4326-8062-92d68899d7e4/kube-rbac-proxy/0.log" Dec 09 11:18:43 crc kubenswrapper[4824]: I1209 11:18:43.669416 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-bf7pj_ccc192e8-2648-4b1b-8420-1cbd1a27a916/kube-rbac-proxy/0.log" Dec 09 11:18:43 crc kubenswrapper[4824]: I1209 11:18:43.747376 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-vx285_54693584-08f2-4326-8062-92d68899d7e4/manager/0.log" Dec 09 11:18:43 crc kubenswrapper[4824]: I1209 11:18:43.907812 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-bf7pj_ccc192e8-2648-4b1b-8420-1cbd1a27a916/manager/0.log" Dec 09 11:18:43 crc kubenswrapper[4824]: I1209 11:18:43.925879 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:18:43 crc kubenswrapper[4824]: E1209 11:18:43.926231 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=ceilometer-central-agent pod=ceilometer-0_openstack(9339c10c-1cea-4ffc-8931-9b349a1e74cd)\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:18:43 crc kubenswrapper[4824]: I1209 11:18:43.954244 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-k92kx_30f20f09-f33f-4c7a-a4b4-41b51f73d692/kube-rbac-proxy/0.log" Dec 09 11:18:43 crc kubenswrapper[4824]: I1209 11:18:43.990530 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-k92kx_30f20f09-f33f-4c7a-a4b4-41b51f73d692/manager/0.log" Dec 09 11:18:44 crc kubenswrapper[4824]: I1209 11:18:44.148098 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-p5zwg_1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97/kube-rbac-proxy/0.log" Dec 09 11:18:44 crc kubenswrapper[4824]: I1209 11:18:44.394843 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-lfprd_b0c20e13-b233-4b21-8f47-facef9603735/kube-rbac-proxy/0.log" Dec 09 11:18:44 crc kubenswrapper[4824]: I1209 11:18:44.424192 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-lfprd_b0c20e13-b233-4b21-8f47-facef9603735/manager/1.log" Dec 09 11:18:44 crc kubenswrapper[4824]: I1209 11:18:44.479808 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-lfprd_b0c20e13-b233-4b21-8f47-facef9603735/manager/0.log" Dec 09 11:18:44 crc kubenswrapper[4824]: I1209 11:18:44.489767 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-p5zwg_1041e8ed-5d4a-4d51-82fb-b1d9f02e5e97/manager/0.log" Dec 09 11:18:44 crc kubenswrapper[4824]: I1209 11:18:44.602711 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-4q2h7_7de1c7e2-adcb-4311-926c-ffc42e3e5fde/kube-rbac-proxy/0.log" Dec 09 11:18:44 crc kubenswrapper[4824]: I1209 11:18:44.726598 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-4q2h7_7de1c7e2-adcb-4311-926c-ffc42e3e5fde/manager/1.log" Dec 09 11:18:44 crc kubenswrapper[4824]: I1209 11:18:44.805133 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-4q2h7_7de1c7e2-adcb-4311-926c-ffc42e3e5fde/manager/0.log" Dec 09 11:18:44 crc kubenswrapper[4824]: I1209 11:18:44.895397 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-9z2kx_d1323625-eb0f-4606-bee7-7acbd453d4a5/kube-rbac-proxy/0.log" Dec 09 11:18:45 crc kubenswrapper[4824]: I1209 11:18:45.621899 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-9z2kx_d1323625-eb0f-4606-bee7-7acbd453d4a5/manager/0.log" Dec 09 11:18:45 crc kubenswrapper[4824]: I1209 11:18:45.810437 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-r9fcl_b6577050-6a21-43bb-84aa-20aad247aafc/kube-rbac-proxy/0.log" Dec 09 11:18:45 crc kubenswrapper[4824]: I1209 11:18:45.827663 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-r9fcl_b6577050-6a21-43bb-84aa-20aad247aafc/manager/1.log" Dec 09 11:18:45 crc kubenswrapper[4824]: I1209 11:18:45.962824 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-r9fcl_b6577050-6a21-43bb-84aa-20aad247aafc/manager/0.log" Dec 09 11:18:46 crc kubenswrapper[4824]: I1209 11:18:46.063977 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-hlwv8_23621a98-1d6b-421a-976c-965cecfe6db1/kube-rbac-proxy/0.log" Dec 09 11:18:46 crc kubenswrapper[4824]: I1209 11:18:46.074288 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-hlwv8_23621a98-1d6b-421a-976c-965cecfe6db1/manager/1.log" Dec 09 11:18:46 crc kubenswrapper[4824]: I1209 11:18:46.210391 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-hlwv8_23621a98-1d6b-421a-976c-965cecfe6db1/manager/0.log" Dec 09 11:18:46 crc kubenswrapper[4824]: I1209 11:18:46.285483 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-rbwkh_d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e/kube-rbac-proxy/0.log" Dec 09 11:18:46 crc kubenswrapper[4824]: I1209 11:18:46.360727 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-rbwkh_d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e/manager/1.log" Dec 09 11:18:46 crc kubenswrapper[4824]: I1209 11:18:46.510625 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-5vnrh_91bf6cd9-a4fe-46e8-b1e7-39e57c1398b3/kube-rbac-proxy/0.log" Dec 09 11:18:46 crc kubenswrapper[4824]: I1209 11:18:46.516618 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-rbwkh_d702a34a-ed07-4d09-a72b-8cbcfe2c1d6e/manager/0.log" Dec 09 11:18:46 crc kubenswrapper[4824]: I1209 11:18:46.551684 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-5vnrh_91bf6cd9-a4fe-46e8-b1e7-39e57c1398b3/manager/1.log" Dec 09 11:18:46 crc kubenswrapper[4824]: I1209 11:18:46.570290 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-5vnrh_91bf6cd9-a4fe-46e8-b1e7-39e57c1398b3/manager/0.log" Dec 09 11:18:46 crc kubenswrapper[4824]: I1209 11:18:46.692215 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879ftr284_c2889de3-49b2-4465-8235-37f9e58b42a3/manager/1.log" Dec 09 11:18:46 crc kubenswrapper[4824]: I1209 11:18:46.738547 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879ftr284_c2889de3-49b2-4465-8235-37f9e58b42a3/kube-rbac-proxy/0.log" Dec 09 11:18:46 crc kubenswrapper[4824]: I1209 11:18:46.765477 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879ftr284_c2889de3-49b2-4465-8235-37f9e58b42a3/manager/0.log" Dec 09 11:18:47 crc kubenswrapper[4824]: I1209 11:18:47.059772 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-2jnrg_25c6995d-6a34-43fc-9b1b-6c34ed9aec46/registry-server/0.log" Dec 09 11:18:47 crc kubenswrapper[4824]: I1209 11:18:47.197457 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-b846b8d48-dhtnn_ebb1c8e1-3960-4c38-969b-c483143c87aa/operator/0.log" Dec 09 11:18:47 crc kubenswrapper[4824]: I1209 11:18:47.913106 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-z68l9_31749799-9eb4-403f-a61d-9d50d8bc8367/kube-rbac-proxy/0.log" Dec 09 11:18:47 crc kubenswrapper[4824]: I1209 11:18:47.948321 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-z68l9_31749799-9eb4-403f-a61d-9d50d8bc8367/manager/0.log" Dec 09 11:18:48 crc kubenswrapper[4824]: I1209 11:18:48.087666 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-pqhrq_383a67da-0f83-41e2-82d6-809d1b6cdbec/manager/1.log" Dec 09 11:18:48 crc kubenswrapper[4824]: I1209 11:18:48.187577 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-pqhrq_383a67da-0f83-41e2-82d6-809d1b6cdbec/kube-rbac-proxy/0.log" Dec 09 11:18:48 crc kubenswrapper[4824]: I1209 11:18:48.213991 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-pqhrq_383a67da-0f83-41e2-82d6-809d1b6cdbec/manager/0.log" Dec 09 11:18:48 crc kubenswrapper[4824]: I1209 11:18:48.414756 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-lwtpc_09d4392b-616f-4531-837e-46f1ef396caf/operator/0.log" Dec 09 11:18:48 crc kubenswrapper[4824]: I1209 11:18:48.454758 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-rz4sl_6e91c44f-6221-4dfd-9c18-2bbedff02850/kube-rbac-proxy/0.log" Dec 09 11:18:48 crc kubenswrapper[4824]: I1209 11:18:48.460932 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-rz4sl_6e91c44f-6221-4dfd-9c18-2bbedff02850/manager/1.log" Dec 09 11:18:48 crc kubenswrapper[4824]: I1209 11:18:48.629032 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-rz4sl_6e91c44f-6221-4dfd-9c18-2bbedff02850/manager/0.log" Dec 09 11:18:48 crc kubenswrapper[4824]: I1209 11:18:48.750279 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-577c7bd649-xtnkl_feab57e6-788b-4f23-9e99-aa248786052d/manager/1.log" Dec 09 11:18:48 crc kubenswrapper[4824]: I1209 11:18:48.781482 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-577c7bd649-xtnkl_feab57e6-788b-4f23-9e99-aa248786052d/kube-rbac-proxy/0.log" Dec 09 11:18:48 crc kubenswrapper[4824]: I1209 11:18:48.824206 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-598df84fd8-l2bt7_e7e496dc-23c4-47cd-9c5a-aa4430130849/manager/0.log" Dec 09 11:18:49 crc kubenswrapper[4824]: I1209 11:18:49.025074 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-55kwm_bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8/kube-rbac-proxy/0.log" Dec 09 11:18:49 crc kubenswrapper[4824]: I1209 11:18:49.069399 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-55kwm_bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8/manager/1.log" Dec 09 11:18:49 crc kubenswrapper[4824]: I1209 11:18:49.160371 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-577c7bd649-xtnkl_feab57e6-788b-4f23-9e99-aa248786052d/manager/0.log" Dec 09 11:18:49 crc kubenswrapper[4824]: I1209 11:18:49.178487 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-55kwm_bf3307d5-cd8f-4184-b0e9-d2d1e26b73e8/manager/0.log" Dec 09 11:18:49 crc kubenswrapper[4824]: I1209 11:18:49.239542 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-q7psb_0af0249c-ed71-4f92-b1a3-cb6f43b92529/kube-rbac-proxy/0.log" Dec 09 11:18:49 crc kubenswrapper[4824]: I1209 11:18:49.371056 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-q7psb_0af0249c-ed71-4f92-b1a3-cb6f43b92529/manager/1.log" Dec 09 11:18:49 crc kubenswrapper[4824]: I1209 11:18:49.394556 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-q7psb_0af0249c-ed71-4f92-b1a3-cb6f43b92529/manager/0.log" Dec 09 11:18:55 crc kubenswrapper[4824]: I1209 11:18:55.911407 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:18:55 crc kubenswrapper[4824]: E1209 11:18:55.912246 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:18:56 crc kubenswrapper[4824]: I1209 11:18:56.912065 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:18:56 crc kubenswrapper[4824]: E1209 11:18:56.912840 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=ceilometer-central-agent pod=ceilometer-0_openstack(9339c10c-1cea-4ffc-8931-9b349a1e74cd)\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:19:08 crc kubenswrapper[4824]: I1209 11:19:08.912949 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:19:08 crc kubenswrapper[4824]: E1209 11:19:08.913857 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:19:09 crc kubenswrapper[4824]: I1209 11:19:09.868803 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-9jtv8_43d3560f-ad7b-4037-a980-0ed995e7e1d2/control-plane-machine-set-operator/0.log" Dec 09 11:19:10 crc kubenswrapper[4824]: I1209 11:19:10.086100 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-4rv7m_14ebfe85-ea7c-49a4-a2e2-f8822c3a111e/kube-rbac-proxy/0.log" Dec 09 11:19:10 crc kubenswrapper[4824]: I1209 11:19:10.114556 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-4rv7m_14ebfe85-ea7c-49a4-a2e2-f8822c3a111e/machine-api-operator/0.log" Dec 09 11:19:11 crc kubenswrapper[4824]: I1209 11:19:11.918573 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:19:11 crc kubenswrapper[4824]: E1209 11:19:11.919168 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=ceilometer-central-agent pod=ceilometer-0_openstack(9339c10c-1cea-4ffc-8931-9b349a1e74cd)\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:19:21 crc kubenswrapper[4824]: I1209 11:19:21.911117 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:19:21 crc kubenswrapper[4824]: E1209 11:19:21.912015 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:19:22 crc kubenswrapper[4824]: I1209 11:19:22.712079 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-2zl4f_a3e087c0-4d4b-4678-8a71-1cf7ee33a867/cert-manager-controller/0.log" Dec 09 11:19:22 crc kubenswrapper[4824]: I1209 11:19:22.911900 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:19:22 crc kubenswrapper[4824]: E1209 11:19:22.912245 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=ceilometer-central-agent pod=ceilometer-0_openstack(9339c10c-1cea-4ffc-8931-9b349a1e74cd)\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:19:22 crc kubenswrapper[4824]: I1209 11:19:22.920243 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-d4569_c35e598c-4ffa-4c4b-bdec-9940fe209d32/cert-manager-cainjector/0.log" Dec 09 11:19:22 crc kubenswrapper[4824]: I1209 11:19:22.971979 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-dbnmx_9f87669d-33a9-4269-808c-ca6c718f762c/cert-manager-webhook/1.log" Dec 09 11:19:22 crc kubenswrapper[4824]: I1209 11:19:22.991267 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-dbnmx_9f87669d-33a9-4269-808c-ca6c718f762c/cert-manager-webhook/0.log" Dec 09 11:19:36 crc kubenswrapper[4824]: I1209 11:19:36.910944 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:19:36 crc kubenswrapper[4824]: I1209 11:19:36.911600 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:19:36 crc kubenswrapper[4824]: E1209 11:19:36.911907 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:19:36 crc kubenswrapper[4824]: I1209 11:19:36.930093 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 11:19:37 crc kubenswrapper[4824]: E1209 11:19:37.045518 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 11:19:37 crc kubenswrapper[4824]: E1209 11:19:37.045596 4824 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 11:19:37 crc kubenswrapper[4824]: E1209 11:19:37.046411 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n9h55h68ch65bh655h64chbfh6ch5b4h575h6bh654h5bdh587h5h58bh5dch5b9hf4h548h579hf8h5f5h686h667h97h57hb5h65h656h588h5f8q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6p7fx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9339c10c-1cea-4ffc-8931-9b349a1e74cd): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 11:19:37 crc kubenswrapper[4824]: E1209 11:19:37.047702 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:19:37 crc kubenswrapper[4824]: I1209 11:19:37.180948 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-w6n8z_d9469d6d-b886-4086-8c80-9ea36f65574d/nmstate-console-plugin/0.log" Dec 09 11:19:37 crc kubenswrapper[4824]: I1209 11:19:37.323237 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-5strj_952171d0-1a2e-4801-bc11-4012ba19588c/nmstate-handler/0.log" Dec 09 11:19:37 crc kubenswrapper[4824]: I1209 11:19:37.323483 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-z8zrc_1927c56b-adc3-4088-8d46-8cb3cf9acc1c/kube-rbac-proxy/0.log" Dec 09 11:19:37 crc kubenswrapper[4824]: I1209 11:19:37.433874 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-z8zrc_1927c56b-adc3-4088-8d46-8cb3cf9acc1c/nmstate-metrics/0.log" Dec 09 11:19:37 crc kubenswrapper[4824]: I1209 11:19:37.567907 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-45dxs_d9389a26-39b2-43f3-bd7b-54f83b8060ed/nmstate-operator/0.log" Dec 09 11:19:37 crc kubenswrapper[4824]: I1209 11:19:37.617235 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-jklpr_0b932274-5eba-4e0d-8a64-1f469a6ab3d1/nmstate-webhook/0.log" Dec 09 11:19:50 crc kubenswrapper[4824]: I1209 11:19:50.680330 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6bf69c47b7-8fjw7_352c7865-5a0b-4ebb-93f2-513ce433ff8f/manager/1.log" Dec 09 11:19:50 crc kubenswrapper[4824]: I1209 11:19:50.731378 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6bf69c47b7-8fjw7_352c7865-5a0b-4ebb-93f2-513ce433ff8f/kube-rbac-proxy/0.log" Dec 09 11:19:50 crc kubenswrapper[4824]: I1209 11:19:50.911157 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:19:50 crc kubenswrapper[4824]: E1209 11:19:50.911591 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:19:51 crc kubenswrapper[4824]: I1209 11:19:51.540080 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6bf69c47b7-8fjw7_352c7865-5a0b-4ebb-93f2-513ce433ff8f/manager/0.log" Dec 09 11:19:51 crc kubenswrapper[4824]: I1209 11:19:51.914014 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:19:51 crc kubenswrapper[4824]: E1209 11:19:51.917254 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:20:04 crc kubenswrapper[4824]: I1209 11:20:04.299844 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:20:04 crc kubenswrapper[4824]: E1209 11:20:04.300607 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:20:06 crc kubenswrapper[4824]: I1209 11:20:06.911437 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:20:06 crc kubenswrapper[4824]: E1209 11:20:06.914293 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:20:07 crc kubenswrapper[4824]: I1209 11:20:07.782344 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_cluster-logging-operator-ff9846bd-txx2m_333c3894-c96b-4511-8c1b-8ea76bae7ef6/cluster-logging-operator/0.log" Dec 09 11:20:07 crc kubenswrapper[4824]: I1209 11:20:07.896073 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_collector-v9zmd_91d19c8e-6e12-4dc2-a7bc-aa256d10fabe/collector/0.log" Dec 09 11:20:07 crc kubenswrapper[4824]: I1209 11:20:07.994561 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-compactor-0_379b9136-1cdf-4786-8494-e99c7161b202/loki-compactor/0.log" Dec 09 11:20:08 crc kubenswrapper[4824]: I1209 11:20:08.093818 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-distributor-76cc67bf56-nl2vt_72ff7b94-63fd-45d1-a803-bc1ccf4388ed/loki-distributor/0.log" Dec 09 11:20:08 crc kubenswrapper[4824]: I1209 11:20:08.225229 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-b549956cc-bvp6v_858d899b-800f-4639-8fbd-4f1ccad44991/gateway/0.log" Dec 09 11:20:08 crc kubenswrapper[4824]: I1209 11:20:08.273297 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-b549956cc-bvp6v_858d899b-800f-4639-8fbd-4f1ccad44991/opa/0.log" Dec 09 11:20:08 crc kubenswrapper[4824]: I1209 11:20:08.413296 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-b549956cc-vz44w_93c989b8-7b86-4339-bbab-5886c7d13dc9/gateway/0.log" Dec 09 11:20:08 crc kubenswrapper[4824]: I1209 11:20:08.474601 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-b549956cc-vz44w_93c989b8-7b86-4339-bbab-5886c7d13dc9/opa/0.log" Dec 09 11:20:08 crc kubenswrapper[4824]: I1209 11:20:08.530119 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-index-gateway-0_6baf0bfe-7be6-4309-93da-e86174e4654e/loki-index-gateway/0.log" Dec 09 11:20:08 crc kubenswrapper[4824]: I1209 11:20:08.788121 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-ingester-0_9d24f296-729f-4eec-a1ae-d6b904399394/loki-ingester/0.log" Dec 09 11:20:08 crc kubenswrapper[4824]: I1209 11:20:08.803076 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-querier-5895d59bb8-89dgf_1006b7c9-7244-44af-8bc9-52787d891f7f/loki-querier/0.log" Dec 09 11:20:08 crc kubenswrapper[4824]: I1209 11:20:08.959261 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-query-frontend-84558f7c9f-9d5fm_e05c11bd-cd92-4bac-adea-0a6049ccfb39/loki-query-frontend/0.log" Dec 09 11:20:17 crc kubenswrapper[4824]: I1209 11:20:17.911523 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:20:17 crc kubenswrapper[4824]: E1209 11:20:17.914883 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:20:18 crc kubenswrapper[4824]: I1209 11:20:18.910722 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:20:18 crc kubenswrapper[4824]: E1209 11:20:18.911275 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:20:25 crc kubenswrapper[4824]: I1209 11:20:25.306645 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-wxgt4_3af18cd9-4124-4f24-83ea-0ad26ad9ae43/controller/1.log" Dec 09 11:20:25 crc kubenswrapper[4824]: I1209 11:20:25.455655 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-wxgt4_3af18cd9-4124-4f24-83ea-0ad26ad9ae43/controller/0.log" Dec 09 11:20:25 crc kubenswrapper[4824]: I1209 11:20:25.536966 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-wxgt4_3af18cd9-4124-4f24-83ea-0ad26ad9ae43/kube-rbac-proxy/0.log" Dec 09 11:20:25 crc kubenswrapper[4824]: I1209 11:20:25.585684 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntx24_ad8f524a-8c95-47ab-b74d-9f83331fce76/cp-frr-files/0.log" Dec 09 11:20:25 crc kubenswrapper[4824]: I1209 11:20:25.817586 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntx24_ad8f524a-8c95-47ab-b74d-9f83331fce76/cp-frr-files/0.log" Dec 09 11:20:25 crc kubenswrapper[4824]: I1209 11:20:25.831266 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntx24_ad8f524a-8c95-47ab-b74d-9f83331fce76/cp-metrics/0.log" Dec 09 11:20:25 crc kubenswrapper[4824]: I1209 11:20:25.838521 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntx24_ad8f524a-8c95-47ab-b74d-9f83331fce76/cp-reloader/0.log" Dec 09 11:20:25 crc kubenswrapper[4824]: I1209 11:20:25.864065 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntx24_ad8f524a-8c95-47ab-b74d-9f83331fce76/cp-reloader/0.log" Dec 09 11:20:26 crc kubenswrapper[4824]: I1209 11:20:26.044332 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntx24_ad8f524a-8c95-47ab-b74d-9f83331fce76/cp-metrics/0.log" Dec 09 11:20:26 crc kubenswrapper[4824]: I1209 11:20:26.080803 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntx24_ad8f524a-8c95-47ab-b74d-9f83331fce76/cp-metrics/0.log" Dec 09 11:20:26 crc kubenswrapper[4824]: I1209 11:20:26.087174 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntx24_ad8f524a-8c95-47ab-b74d-9f83331fce76/cp-frr-files/0.log" Dec 09 11:20:26 crc kubenswrapper[4824]: I1209 11:20:26.090093 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntx24_ad8f524a-8c95-47ab-b74d-9f83331fce76/cp-reloader/0.log" Dec 09 11:20:26 crc kubenswrapper[4824]: I1209 11:20:26.286300 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntx24_ad8f524a-8c95-47ab-b74d-9f83331fce76/cp-reloader/0.log" Dec 09 11:20:26 crc kubenswrapper[4824]: I1209 11:20:26.286357 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntx24_ad8f524a-8c95-47ab-b74d-9f83331fce76/cp-frr-files/0.log" Dec 09 11:20:26 crc kubenswrapper[4824]: I1209 11:20:26.288344 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntx24_ad8f524a-8c95-47ab-b74d-9f83331fce76/cp-metrics/0.log" Dec 09 11:20:26 crc kubenswrapper[4824]: I1209 11:20:26.322270 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntx24_ad8f524a-8c95-47ab-b74d-9f83331fce76/controller/0.log" Dec 09 11:20:26 crc kubenswrapper[4824]: I1209 11:20:26.528502 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntx24_ad8f524a-8c95-47ab-b74d-9f83331fce76/frr/1.log" Dec 09 11:20:26 crc kubenswrapper[4824]: I1209 11:20:26.529242 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntx24_ad8f524a-8c95-47ab-b74d-9f83331fce76/kube-rbac-proxy/0.log" Dec 09 11:20:26 crc kubenswrapper[4824]: I1209 11:20:26.563705 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntx24_ad8f524a-8c95-47ab-b74d-9f83331fce76/frr-metrics/0.log" Dec 09 11:20:27 crc kubenswrapper[4824]: I1209 11:20:27.463237 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntx24_ad8f524a-8c95-47ab-b74d-9f83331fce76/kube-rbac-proxy-frr/0.log" Dec 09 11:20:27 crc kubenswrapper[4824]: I1209 11:20:27.487059 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntx24_ad8f524a-8c95-47ab-b74d-9f83331fce76/reloader/0.log" Dec 09 11:20:27 crc kubenswrapper[4824]: I1209 11:20:27.713850 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-8rvmd_bd655e5b-2ac3-4fe8-ad58-6012d0180897/frr-k8s-webhook-server/0.log" Dec 09 11:20:27 crc kubenswrapper[4824]: I1209 11:20:27.740281 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-69bb78d57c-p4l9h_2fed244b-a135-4d95-a5e5-2eea4bcde7cb/manager/0.log" Dec 09 11:20:28 crc kubenswrapper[4824]: I1209 11:20:28.030096 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5dcb7584cb-dfrlw_bab491e7-4eec-4ba1-975d-fb2468372ade/webhook-server/1.log" Dec 09 11:20:28 crc kubenswrapper[4824]: I1209 11:20:28.125614 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5dcb7584cb-dfrlw_bab491e7-4eec-4ba1-975d-fb2468372ade/webhook-server/0.log" Dec 09 11:20:28 crc kubenswrapper[4824]: I1209 11:20:28.274198 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ntx24_ad8f524a-8c95-47ab-b74d-9f83331fce76/frr/0.log" Dec 09 11:20:28 crc kubenswrapper[4824]: I1209 11:20:28.566820 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5xfv2_87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe/kube-rbac-proxy/0.log" Dec 09 11:20:28 crc kubenswrapper[4824]: I1209 11:20:28.807984 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5xfv2_87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe/speaker/1.log" Dec 09 11:20:29 crc kubenswrapper[4824]: I1209 11:20:29.224662 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5xfv2_87dca6f6-ef3a-4fc2-a4e2-49f5a54257fe/speaker/0.log" Dec 09 11:20:31 crc kubenswrapper[4824]: I1209 11:20:31.912133 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:20:31 crc kubenswrapper[4824]: I1209 11:20:31.912701 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:20:31 crc kubenswrapper[4824]: E1209 11:20:31.912986 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:20:31 crc kubenswrapper[4824]: E1209 11:20:31.914963 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:20:43 crc kubenswrapper[4824]: I1209 11:20:43.762255 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5_cf722168-25bd-4faa-9d35-b59240bb718f/util/0.log" Dec 09 11:20:43 crc kubenswrapper[4824]: I1209 11:20:43.928925 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:20:43 crc kubenswrapper[4824]: E1209 11:20:43.929330 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:20:43 crc kubenswrapper[4824]: I1209 11:20:43.990562 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5_cf722168-25bd-4faa-9d35-b59240bb718f/util/0.log" Dec 09 11:20:44 crc kubenswrapper[4824]: I1209 11:20:44.052950 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5_cf722168-25bd-4faa-9d35-b59240bb718f/pull/0.log" Dec 09 11:20:44 crc kubenswrapper[4824]: I1209 11:20:44.069899 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5_cf722168-25bd-4faa-9d35-b59240bb718f/pull/0.log" Dec 09 11:20:44 crc kubenswrapper[4824]: I1209 11:20:44.241047 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5_cf722168-25bd-4faa-9d35-b59240bb718f/util/0.log" Dec 09 11:20:44 crc kubenswrapper[4824]: I1209 11:20:44.283427 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5_cf722168-25bd-4faa-9d35-b59240bb718f/pull/0.log" Dec 09 11:20:44 crc kubenswrapper[4824]: I1209 11:20:44.287628 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8gwnr5_cf722168-25bd-4faa-9d35-b59240bb718f/extract/0.log" Dec 09 11:20:44 crc kubenswrapper[4824]: I1209 11:20:44.487905 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc_42a87103-eb40-4e22-97d7-18ff2885d50b/util/0.log" Dec 09 11:20:44 crc kubenswrapper[4824]: I1209 11:20:44.702293 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc_42a87103-eb40-4e22-97d7-18ff2885d50b/pull/0.log" Dec 09 11:20:44 crc kubenswrapper[4824]: I1209 11:20:44.733259 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc_42a87103-eb40-4e22-97d7-18ff2885d50b/util/0.log" Dec 09 11:20:44 crc kubenswrapper[4824]: I1209 11:20:44.763194 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc_42a87103-eb40-4e22-97d7-18ff2885d50b/pull/0.log" Dec 09 11:20:44 crc kubenswrapper[4824]: I1209 11:20:44.911169 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:20:44 crc kubenswrapper[4824]: E1209 11:20:44.914535 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:20:45 crc kubenswrapper[4824]: I1209 11:20:45.040507 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc_42a87103-eb40-4e22-97d7-18ff2885d50b/util/0.log" Dec 09 11:20:45 crc kubenswrapper[4824]: I1209 11:20:45.103727 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc_42a87103-eb40-4e22-97d7-18ff2885d50b/pull/0.log" Dec 09 11:20:45 crc kubenswrapper[4824]: I1209 11:20:45.137202 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqwfnc_42a87103-eb40-4e22-97d7-18ff2885d50b/extract/0.log" Dec 09 11:20:45 crc kubenswrapper[4824]: I1209 11:20:45.544758 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8_d28511a4-2458-434b-85ad-98d2253d4d7b/util/0.log" Dec 09 11:20:45 crc kubenswrapper[4824]: I1209 11:20:45.764442 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8_d28511a4-2458-434b-85ad-98d2253d4d7b/util/0.log" Dec 09 11:20:45 crc kubenswrapper[4824]: I1209 11:20:45.785349 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8_d28511a4-2458-434b-85ad-98d2253d4d7b/pull/0.log" Dec 09 11:20:45 crc kubenswrapper[4824]: I1209 11:20:45.816266 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8_d28511a4-2458-434b-85ad-98d2253d4d7b/pull/0.log" Dec 09 11:20:46 crc kubenswrapper[4824]: I1209 11:20:46.002940 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8_d28511a4-2458-434b-85ad-98d2253d4d7b/pull/0.log" Dec 09 11:20:46 crc kubenswrapper[4824]: I1209 11:20:46.026839 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8_d28511a4-2458-434b-85ad-98d2253d4d7b/extract/0.log" Dec 09 11:20:46 crc kubenswrapper[4824]: I1209 11:20:46.044347 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210pcfk8_d28511a4-2458-434b-85ad-98d2253d4d7b/util/0.log" Dec 09 11:20:46 crc kubenswrapper[4824]: I1209 11:20:46.296250 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz_bf54b053-f1e7-4b01-90f8-2c0aac92c4b4/util/0.log" Dec 09 11:20:46 crc kubenswrapper[4824]: I1209 11:20:46.510711 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz_bf54b053-f1e7-4b01-90f8-2c0aac92c4b4/util/0.log" Dec 09 11:20:46 crc kubenswrapper[4824]: I1209 11:20:46.542816 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz_bf54b053-f1e7-4b01-90f8-2c0aac92c4b4/pull/0.log" Dec 09 11:20:46 crc kubenswrapper[4824]: I1209 11:20:46.573966 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz_bf54b053-f1e7-4b01-90f8-2c0aac92c4b4/pull/0.log" Dec 09 11:20:46 crc kubenswrapper[4824]: I1209 11:20:46.735940 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz_bf54b053-f1e7-4b01-90f8-2c0aac92c4b4/util/0.log" Dec 09 11:20:46 crc kubenswrapper[4824]: I1209 11:20:46.736922 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz_bf54b053-f1e7-4b01-90f8-2c0aac92c4b4/extract/0.log" Dec 09 11:20:46 crc kubenswrapper[4824]: I1209 11:20:46.742899 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fkqmvz_bf54b053-f1e7-4b01-90f8-2c0aac92c4b4/pull/0.log" Dec 09 11:20:46 crc kubenswrapper[4824]: I1209 11:20:46.946306 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t_44aedeb1-3d60-43f0-9a8b-472e489077e0/util/0.log" Dec 09 11:20:47 crc kubenswrapper[4824]: I1209 11:20:47.205997 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t_44aedeb1-3d60-43f0-9a8b-472e489077e0/pull/0.log" Dec 09 11:20:47 crc kubenswrapper[4824]: I1209 11:20:47.246019 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t_44aedeb1-3d60-43f0-9a8b-472e489077e0/pull/0.log" Dec 09 11:20:47 crc kubenswrapper[4824]: I1209 11:20:47.253781 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t_44aedeb1-3d60-43f0-9a8b-472e489077e0/util/0.log" Dec 09 11:20:47 crc kubenswrapper[4824]: I1209 11:20:47.453939 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t_44aedeb1-3d60-43f0-9a8b-472e489077e0/util/0.log" Dec 09 11:20:47 crc kubenswrapper[4824]: I1209 11:20:47.467474 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t_44aedeb1-3d60-43f0-9a8b-472e489077e0/pull/0.log" Dec 09 11:20:47 crc kubenswrapper[4824]: I1209 11:20:47.484566 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834j68t_44aedeb1-3d60-43f0-9a8b-472e489077e0/extract/0.log" Dec 09 11:20:47 crc kubenswrapper[4824]: I1209 11:20:47.670659 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2cbmd_257d6d77-75f1-41e0-8bd2-c6ec3c1ff582/extract-utilities/0.log" Dec 09 11:20:47 crc kubenswrapper[4824]: I1209 11:20:47.902085 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2cbmd_257d6d77-75f1-41e0-8bd2-c6ec3c1ff582/extract-utilities/0.log" Dec 09 11:20:47 crc kubenswrapper[4824]: I1209 11:20:47.921590 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2cbmd_257d6d77-75f1-41e0-8bd2-c6ec3c1ff582/extract-content/0.log" Dec 09 11:20:47 crc kubenswrapper[4824]: I1209 11:20:47.965684 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2cbmd_257d6d77-75f1-41e0-8bd2-c6ec3c1ff582/extract-content/0.log" Dec 09 11:20:48 crc kubenswrapper[4824]: I1209 11:20:48.172895 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2cbmd_257d6d77-75f1-41e0-8bd2-c6ec3c1ff582/extract-content/0.log" Dec 09 11:20:48 crc kubenswrapper[4824]: I1209 11:20:48.187940 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2cbmd_257d6d77-75f1-41e0-8bd2-c6ec3c1ff582/extract-utilities/0.log" Dec 09 11:20:48 crc kubenswrapper[4824]: I1209 11:20:48.345558 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2cbmd_257d6d77-75f1-41e0-8bd2-c6ec3c1ff582/registry-server/1.log" Dec 09 11:20:48 crc kubenswrapper[4824]: I1209 11:20:48.471808 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jm5ld_bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271/extract-utilities/0.log" Dec 09 11:20:48 crc kubenswrapper[4824]: I1209 11:20:48.730970 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jm5ld_bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271/extract-utilities/0.log" Dec 09 11:20:48 crc kubenswrapper[4824]: I1209 11:20:48.738347 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jm5ld_bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271/extract-content/0.log" Dec 09 11:20:48 crc kubenswrapper[4824]: I1209 11:20:48.928358 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jm5ld_bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271/extract-content/0.log" Dec 09 11:20:48 crc kubenswrapper[4824]: I1209 11:20:48.952423 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2cbmd_257d6d77-75f1-41e0-8bd2-c6ec3c1ff582/registry-server/0.log" Dec 09 11:20:49 crc kubenswrapper[4824]: I1209 11:20:49.009926 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jm5ld_bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271/extract-content/0.log" Dec 09 11:20:49 crc kubenswrapper[4824]: I1209 11:20:49.058988 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jm5ld_bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271/extract-utilities/0.log" Dec 09 11:20:49 crc kubenswrapper[4824]: I1209 11:20:49.196467 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jm5ld_bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271/registry-server/1.log" Dec 09 11:20:49 crc kubenswrapper[4824]: I1209 11:20:49.515800 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-95jfr_6ffd99b7-9bb8-479e-a1a1-7f31ee006e38/extract-utilities/0.log" Dec 09 11:20:49 crc kubenswrapper[4824]: I1209 11:20:49.543773 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4tt6w_1b70aae4-4b6b-457c-a358-0f7ce11b8206/marketplace-operator/0.log" Dec 09 11:20:50 crc kubenswrapper[4824]: I1209 11:20:50.168570 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jm5ld_bb0311bb-7c1c-43a6-aac3-8d6dcdcf6271/registry-server/0.log" Dec 09 11:20:50 crc kubenswrapper[4824]: I1209 11:20:50.762612 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-95jfr_6ffd99b7-9bb8-479e-a1a1-7f31ee006e38/extract-content/0.log" Dec 09 11:20:50 crc kubenswrapper[4824]: I1209 11:20:50.782246 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-95jfr_6ffd99b7-9bb8-479e-a1a1-7f31ee006e38/extract-utilities/0.log" Dec 09 11:20:50 crc kubenswrapper[4824]: I1209 11:20:50.800193 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-95jfr_6ffd99b7-9bb8-479e-a1a1-7f31ee006e38/extract-content/0.log" Dec 09 11:20:51 crc kubenswrapper[4824]: I1209 11:20:51.132004 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-95jfr_6ffd99b7-9bb8-479e-a1a1-7f31ee006e38/extract-utilities/0.log" Dec 09 11:20:51 crc kubenswrapper[4824]: I1209 11:20:51.220087 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-95jfr_6ffd99b7-9bb8-479e-a1a1-7f31ee006e38/extract-content/0.log" Dec 09 11:20:51 crc kubenswrapper[4824]: I1209 11:20:51.356612 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-95jfr_6ffd99b7-9bb8-479e-a1a1-7f31ee006e38/registry-server/1.log" Dec 09 11:20:51 crc kubenswrapper[4824]: I1209 11:20:51.360773 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-582xj_f2572cf3-89f1-49d7-9246-9350703e9764/extract-utilities/0.log" Dec 09 11:20:51 crc kubenswrapper[4824]: I1209 11:20:51.416985 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-95jfr_6ffd99b7-9bb8-479e-a1a1-7f31ee006e38/registry-server/0.log" Dec 09 11:20:51 crc kubenswrapper[4824]: I1209 11:20:51.585424 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-582xj_f2572cf3-89f1-49d7-9246-9350703e9764/extract-utilities/0.log" Dec 09 11:20:51 crc kubenswrapper[4824]: I1209 11:20:51.629686 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-582xj_f2572cf3-89f1-49d7-9246-9350703e9764/extract-content/0.log" Dec 09 11:20:51 crc kubenswrapper[4824]: I1209 11:20:51.641939 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-582xj_f2572cf3-89f1-49d7-9246-9350703e9764/extract-content/0.log" Dec 09 11:20:51 crc kubenswrapper[4824]: I1209 11:20:51.864750 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-582xj_f2572cf3-89f1-49d7-9246-9350703e9764/extract-utilities/0.log" Dec 09 11:20:51 crc kubenswrapper[4824]: I1209 11:20:51.867551 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-582xj_f2572cf3-89f1-49d7-9246-9350703e9764/extract-content/0.log" Dec 09 11:20:52 crc kubenswrapper[4824]: I1209 11:20:52.149750 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-582xj_f2572cf3-89f1-49d7-9246-9350703e9764/registry-server/1.log" Dec 09 11:20:52 crc kubenswrapper[4824]: I1209 11:20:52.676795 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-582xj_f2572cf3-89f1-49d7-9246-9350703e9764/registry-server/0.log" Dec 09 11:20:57 crc kubenswrapper[4824]: I1209 11:20:57.911374 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:20:57 crc kubenswrapper[4824]: E1209 11:20:57.912214 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:20:59 crc kubenswrapper[4824]: I1209 11:20:59.910670 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:20:59 crc kubenswrapper[4824]: E1209 11:20:59.913634 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:21:07 crc kubenswrapper[4824]: I1209 11:21:07.443703 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-xb2z8_ca1c7fd5-ccf0-453d-9426-d77e591d729e/prometheus-operator/0.log" Dec 09 11:21:07 crc kubenswrapper[4824]: I1209 11:21:07.668987 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-749b4b8cbb-2mrlf_eb813060-c1d8-433f-be2e-e85d102499d9/prometheus-operator-admission-webhook/0.log" Dec 09 11:21:07 crc kubenswrapper[4824]: I1209 11:21:07.690909 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-749b4b8cbb-9mxxp_cdadfe90-b4b1-4c57-a187-f6e2cbb72103/prometheus-operator-admission-webhook/0.log" Dec 09 11:21:07 crc kubenswrapper[4824]: I1209 11:21:07.895498 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-qn2qz_612ebf47-a6c4-4a62-8d97-91a003f49c44/operator/1.log" Dec 09 11:21:07 crc kubenswrapper[4824]: I1209 11:21:07.935553 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-qn2qz_612ebf47-a6c4-4a62-8d97-91a003f49c44/operator/0.log" Dec 09 11:21:08 crc kubenswrapper[4824]: I1209 11:21:08.011323 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-ui-dashboards-7d5fb4cbfb-8j2s2_74341961-8a97-42da-9318-4c33280e45e7/observability-ui-dashboards/0.log" Dec 09 11:21:08 crc kubenswrapper[4824]: I1209 11:21:08.138039 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-pkmss_2ecf9e90-37fe-416d-9e7a-400373cfbc8d/perses-operator/0.log" Dec 09 11:21:11 crc kubenswrapper[4824]: I1209 11:21:11.911412 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:21:13 crc kubenswrapper[4824]: I1209 11:21:13.305615 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerStarted","Data":"49a74dd2910658c46255de301a11501ac6eaebf752b480f0958816296c35a0f8"} Dec 09 11:21:14 crc kubenswrapper[4824]: I1209 11:21:14.910949 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:21:14 crc kubenswrapper[4824]: E1209 11:21:14.913300 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:21:25 crc kubenswrapper[4824]: I1209 11:21:25.435515 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6bf69c47b7-8fjw7_352c7865-5a0b-4ebb-93f2-513ce433ff8f/kube-rbac-proxy/0.log" Dec 09 11:21:25 crc kubenswrapper[4824]: I1209 11:21:25.492659 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6bf69c47b7-8fjw7_352c7865-5a0b-4ebb-93f2-513ce433ff8f/manager/1.log" Dec 09 11:21:25 crc kubenswrapper[4824]: I1209 11:21:25.564047 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6bf69c47b7-8fjw7_352c7865-5a0b-4ebb-93f2-513ce433ff8f/manager/0.log" Dec 09 11:21:26 crc kubenswrapper[4824]: I1209 11:21:26.911842 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:21:26 crc kubenswrapper[4824]: E1209 11:21:26.914833 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:21:39 crc kubenswrapper[4824]: I1209 11:21:39.911060 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:21:39 crc kubenswrapper[4824]: E1209 11:21:39.918271 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:21:50 crc kubenswrapper[4824]: I1209 11:21:50.911399 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:21:50 crc kubenswrapper[4824]: E1209 11:21:50.913833 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:22:02 crc kubenswrapper[4824]: I1209 11:22:02.912621 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:22:02 crc kubenswrapper[4824]: E1209 11:22:02.915675 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:22:17 crc kubenswrapper[4824]: I1209 11:22:17.910992 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:22:18 crc kubenswrapper[4824]: E1209 11:22:18.050890 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 11:22:18 crc kubenswrapper[4824]: E1209 11:22:18.051336 4824 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 11:22:18 crc kubenswrapper[4824]: E1209 11:22:18.051671 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n9h55h68ch65bh655h64chbfh6ch5b4h575h6bh654h5bdh587h5h58bh5dch5b9hf4h548h579hf8h5f5h686h667h97h57hb5h65h656h588h5f8q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6p7fx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9339c10c-1cea-4ffc-8931-9b349a1e74cd): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 11:22:18 crc kubenswrapper[4824]: E1209 11:22:18.053099 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:22:32 crc kubenswrapper[4824]: I1209 11:22:32.030884 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:22:32 crc kubenswrapper[4824]: E1209 11:22:32.036961 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:22:44 crc kubenswrapper[4824]: I1209 11:22:44.910471 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:22:44 crc kubenswrapper[4824]: E1209 11:22:44.914053 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:22:55 crc kubenswrapper[4824]: I1209 11:22:55.911126 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:22:55 crc kubenswrapper[4824]: E1209 11:22:55.914718 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:23:09 crc kubenswrapper[4824]: I1209 11:23:09.911822 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:23:09 crc kubenswrapper[4824]: E1209 11:23:09.914509 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:23:22 crc kubenswrapper[4824]: I1209 11:23:22.911593 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:23:22 crc kubenswrapper[4824]: E1209 11:23:22.915259 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:23:32 crc kubenswrapper[4824]: I1209 11:23:32.861507 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:23:32 crc kubenswrapper[4824]: I1209 11:23:32.862111 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:23:36 crc kubenswrapper[4824]: I1209 11:23:36.911141 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:23:36 crc kubenswrapper[4824]: E1209 11:23:36.915753 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:23:38 crc kubenswrapper[4824]: I1209 11:23:38.657340 4824 generic.go:334] "Generic (PLEG): container finished" podID="594f6c5d-8833-4f51-bc27-52254f02ceaa" containerID="2b92e298c4461b580890011a817668dce792070ce6e3fb399861eb96a5a8993d" exitCode=0 Dec 09 11:23:38 crc kubenswrapper[4824]: I1209 11:23:38.657474 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2w2mq/must-gather-qd8qq" event={"ID":"594f6c5d-8833-4f51-bc27-52254f02ceaa","Type":"ContainerDied","Data":"2b92e298c4461b580890011a817668dce792070ce6e3fb399861eb96a5a8993d"} Dec 09 11:23:38 crc kubenswrapper[4824]: I1209 11:23:38.659154 4824 scope.go:117] "RemoveContainer" containerID="2b92e298c4461b580890011a817668dce792070ce6e3fb399861eb96a5a8993d" Dec 09 11:23:39 crc kubenswrapper[4824]: I1209 11:23:39.517942 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2w2mq_must-gather-qd8qq_594f6c5d-8833-4f51-bc27-52254f02ceaa/gather/0.log" Dec 09 11:23:48 crc kubenswrapper[4824]: I1209 11:23:48.389601 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2w2mq/must-gather-qd8qq"] Dec 09 11:23:48 crc kubenswrapper[4824]: I1209 11:23:48.390383 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-2w2mq/must-gather-qd8qq" podUID="594f6c5d-8833-4f51-bc27-52254f02ceaa" containerName="copy" containerID="cri-o://65752f2c91d1d890d00554d049abfc70baa112c2f7e9d29aac6186ec3b03351d" gracePeriod=2 Dec 09 11:23:48 crc kubenswrapper[4824]: I1209 11:23:48.401497 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2w2mq/must-gather-qd8qq"] Dec 09 11:23:48 crc kubenswrapper[4824]: I1209 11:23:48.797147 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2w2mq_must-gather-qd8qq_594f6c5d-8833-4f51-bc27-52254f02ceaa/copy/0.log" Dec 09 11:23:48 crc kubenswrapper[4824]: I1209 11:23:48.797714 4824 generic.go:334] "Generic (PLEG): container finished" podID="594f6c5d-8833-4f51-bc27-52254f02ceaa" containerID="65752f2c91d1d890d00554d049abfc70baa112c2f7e9d29aac6186ec3b03351d" exitCode=143 Dec 09 11:23:49 crc kubenswrapper[4824]: I1209 11:23:49.369008 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2w2mq_must-gather-qd8qq_594f6c5d-8833-4f51-bc27-52254f02ceaa/copy/0.log" Dec 09 11:23:49 crc kubenswrapper[4824]: I1209 11:23:49.370423 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2w2mq/must-gather-qd8qq" Dec 09 11:23:49 crc kubenswrapper[4824]: I1209 11:23:49.491844 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/594f6c5d-8833-4f51-bc27-52254f02ceaa-must-gather-output\") pod \"594f6c5d-8833-4f51-bc27-52254f02ceaa\" (UID: \"594f6c5d-8833-4f51-bc27-52254f02ceaa\") " Dec 09 11:23:49 crc kubenswrapper[4824]: I1209 11:23:49.492167 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzqbj\" (UniqueName: \"kubernetes.io/projected/594f6c5d-8833-4f51-bc27-52254f02ceaa-kube-api-access-qzqbj\") pod \"594f6c5d-8833-4f51-bc27-52254f02ceaa\" (UID: \"594f6c5d-8833-4f51-bc27-52254f02ceaa\") " Dec 09 11:23:49 crc kubenswrapper[4824]: I1209 11:23:49.507818 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/594f6c5d-8833-4f51-bc27-52254f02ceaa-kube-api-access-qzqbj" (OuterVolumeSpecName: "kube-api-access-qzqbj") pod "594f6c5d-8833-4f51-bc27-52254f02ceaa" (UID: "594f6c5d-8833-4f51-bc27-52254f02ceaa"). InnerVolumeSpecName "kube-api-access-qzqbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:23:49 crc kubenswrapper[4824]: I1209 11:23:49.595167 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzqbj\" (UniqueName: \"kubernetes.io/projected/594f6c5d-8833-4f51-bc27-52254f02ceaa-kube-api-access-qzqbj\") on node \"crc\" DevicePath \"\"" Dec 09 11:23:49 crc kubenswrapper[4824]: I1209 11:23:49.669097 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/594f6c5d-8833-4f51-bc27-52254f02ceaa-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "594f6c5d-8833-4f51-bc27-52254f02ceaa" (UID: "594f6c5d-8833-4f51-bc27-52254f02ceaa"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:23:49 crc kubenswrapper[4824]: I1209 11:23:49.697945 4824 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/594f6c5d-8833-4f51-bc27-52254f02ceaa-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 09 11:23:49 crc kubenswrapper[4824]: I1209 11:23:49.809967 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2w2mq_must-gather-qd8qq_594f6c5d-8833-4f51-bc27-52254f02ceaa/copy/0.log" Dec 09 11:23:49 crc kubenswrapper[4824]: I1209 11:23:49.810698 4824 scope.go:117] "RemoveContainer" containerID="65752f2c91d1d890d00554d049abfc70baa112c2f7e9d29aac6186ec3b03351d" Dec 09 11:23:49 crc kubenswrapper[4824]: I1209 11:23:49.810844 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2w2mq/must-gather-qd8qq" Dec 09 11:23:49 crc kubenswrapper[4824]: I1209 11:23:49.927839 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="594f6c5d-8833-4f51-bc27-52254f02ceaa" path="/var/lib/kubelet/pods/594f6c5d-8833-4f51-bc27-52254f02ceaa/volumes" Dec 09 11:23:50 crc kubenswrapper[4824]: I1209 11:23:50.234157 4824 scope.go:117] "RemoveContainer" containerID="2b92e298c4461b580890011a817668dce792070ce6e3fb399861eb96a5a8993d" Dec 09 11:23:50 crc kubenswrapper[4824]: I1209 11:23:50.911359 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:23:50 crc kubenswrapper[4824]: E1209 11:23:50.913512 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:23:55 crc kubenswrapper[4824]: I1209 11:23:55.039224 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8fpgn"] Dec 09 11:23:55 crc kubenswrapper[4824]: E1209 11:23:55.040577 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="594f6c5d-8833-4f51-bc27-52254f02ceaa" containerName="gather" Dec 09 11:23:55 crc kubenswrapper[4824]: I1209 11:23:55.040596 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="594f6c5d-8833-4f51-bc27-52254f02ceaa" containerName="gather" Dec 09 11:23:55 crc kubenswrapper[4824]: E1209 11:23:55.040645 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d71b52a-7fb2-4c35-89b2-31f237a2d0d1" containerName="container-00" Dec 09 11:23:55 crc kubenswrapper[4824]: I1209 11:23:55.040653 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d71b52a-7fb2-4c35-89b2-31f237a2d0d1" containerName="container-00" Dec 09 11:23:55 crc kubenswrapper[4824]: E1209 11:23:55.040693 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="594f6c5d-8833-4f51-bc27-52254f02ceaa" containerName="copy" Dec 09 11:23:55 crc kubenswrapper[4824]: I1209 11:23:55.040700 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="594f6c5d-8833-4f51-bc27-52254f02ceaa" containerName="copy" Dec 09 11:23:55 crc kubenswrapper[4824]: I1209 11:23:55.041011 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="594f6c5d-8833-4f51-bc27-52254f02ceaa" containerName="copy" Dec 09 11:23:55 crc kubenswrapper[4824]: I1209 11:23:55.041028 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d71b52a-7fb2-4c35-89b2-31f237a2d0d1" containerName="container-00" Dec 09 11:23:55 crc kubenswrapper[4824]: I1209 11:23:55.041068 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="594f6c5d-8833-4f51-bc27-52254f02ceaa" containerName="gather" Dec 09 11:23:55 crc kubenswrapper[4824]: I1209 11:23:55.056230 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8fpgn" Dec 09 11:23:55 crc kubenswrapper[4824]: I1209 11:23:55.128854 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/564a4ea7-ea14-41ee-be38-4eb8c5a22446-catalog-content\") pod \"certified-operators-8fpgn\" (UID: \"564a4ea7-ea14-41ee-be38-4eb8c5a22446\") " pod="openshift-marketplace/certified-operators-8fpgn" Dec 09 11:23:55 crc kubenswrapper[4824]: I1209 11:23:55.129085 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/564a4ea7-ea14-41ee-be38-4eb8c5a22446-utilities\") pod \"certified-operators-8fpgn\" (UID: \"564a4ea7-ea14-41ee-be38-4eb8c5a22446\") " pod="openshift-marketplace/certified-operators-8fpgn" Dec 09 11:23:55 crc kubenswrapper[4824]: I1209 11:23:55.129145 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8fpgn"] Dec 09 11:23:55 crc kubenswrapper[4824]: I1209 11:23:55.129364 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh4d8\" (UniqueName: \"kubernetes.io/projected/564a4ea7-ea14-41ee-be38-4eb8c5a22446-kube-api-access-rh4d8\") pod \"certified-operators-8fpgn\" (UID: \"564a4ea7-ea14-41ee-be38-4eb8c5a22446\") " pod="openshift-marketplace/certified-operators-8fpgn" Dec 09 11:23:55 crc kubenswrapper[4824]: I1209 11:23:55.231106 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/564a4ea7-ea14-41ee-be38-4eb8c5a22446-catalog-content\") pod \"certified-operators-8fpgn\" (UID: \"564a4ea7-ea14-41ee-be38-4eb8c5a22446\") " pod="openshift-marketplace/certified-operators-8fpgn" Dec 09 11:23:55 crc kubenswrapper[4824]: I1209 11:23:55.231199 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/564a4ea7-ea14-41ee-be38-4eb8c5a22446-utilities\") pod \"certified-operators-8fpgn\" (UID: \"564a4ea7-ea14-41ee-be38-4eb8c5a22446\") " pod="openshift-marketplace/certified-operators-8fpgn" Dec 09 11:23:55 crc kubenswrapper[4824]: I1209 11:23:55.231306 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rh4d8\" (UniqueName: \"kubernetes.io/projected/564a4ea7-ea14-41ee-be38-4eb8c5a22446-kube-api-access-rh4d8\") pod \"certified-operators-8fpgn\" (UID: \"564a4ea7-ea14-41ee-be38-4eb8c5a22446\") " pod="openshift-marketplace/certified-operators-8fpgn" Dec 09 11:23:55 crc kubenswrapper[4824]: I1209 11:23:55.231674 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/564a4ea7-ea14-41ee-be38-4eb8c5a22446-catalog-content\") pod \"certified-operators-8fpgn\" (UID: \"564a4ea7-ea14-41ee-be38-4eb8c5a22446\") " pod="openshift-marketplace/certified-operators-8fpgn" Dec 09 11:23:55 crc kubenswrapper[4824]: I1209 11:23:55.231726 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/564a4ea7-ea14-41ee-be38-4eb8c5a22446-utilities\") pod \"certified-operators-8fpgn\" (UID: \"564a4ea7-ea14-41ee-be38-4eb8c5a22446\") " pod="openshift-marketplace/certified-operators-8fpgn" Dec 09 11:23:55 crc kubenswrapper[4824]: I1209 11:23:55.253256 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh4d8\" (UniqueName: \"kubernetes.io/projected/564a4ea7-ea14-41ee-be38-4eb8c5a22446-kube-api-access-rh4d8\") pod \"certified-operators-8fpgn\" (UID: \"564a4ea7-ea14-41ee-be38-4eb8c5a22446\") " pod="openshift-marketplace/certified-operators-8fpgn" Dec 09 11:23:55 crc kubenswrapper[4824]: I1209 11:23:55.382021 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8fpgn" Dec 09 11:23:55 crc kubenswrapper[4824]: I1209 11:23:55.967028 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8fpgn"] Dec 09 11:23:56 crc kubenswrapper[4824]: I1209 11:23:56.921311 4824 generic.go:334] "Generic (PLEG): container finished" podID="564a4ea7-ea14-41ee-be38-4eb8c5a22446" containerID="c9cd33cc3af706051ab8073731d7acb29b927e424a2f81990f7a70ff4bf6f5db" exitCode=0 Dec 09 11:23:56 crc kubenswrapper[4824]: I1209 11:23:56.923595 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8fpgn" event={"ID":"564a4ea7-ea14-41ee-be38-4eb8c5a22446","Type":"ContainerDied","Data":"c9cd33cc3af706051ab8073731d7acb29b927e424a2f81990f7a70ff4bf6f5db"} Dec 09 11:23:56 crc kubenswrapper[4824]: I1209 11:23:56.923725 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8fpgn" event={"ID":"564a4ea7-ea14-41ee-be38-4eb8c5a22446","Type":"ContainerStarted","Data":"341efcfdc037b7c277571b71ed078aae2c032062fafa9e38eb5b8477707e2f99"} Dec 09 11:23:57 crc kubenswrapper[4824]: I1209 11:23:57.934590 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8fpgn" event={"ID":"564a4ea7-ea14-41ee-be38-4eb8c5a22446","Type":"ContainerStarted","Data":"749218e46e0eebdfa29fdae704e00418fac0a74a14d921129a43c057d12f6c1d"} Dec 09 11:23:58 crc kubenswrapper[4824]: I1209 11:23:58.363983 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n7r8x"] Dec 09 11:23:58 crc kubenswrapper[4824]: I1209 11:23:58.367705 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n7r8x" Dec 09 11:23:58 crc kubenswrapper[4824]: I1209 11:23:58.378810 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n7r8x"] Dec 09 11:23:58 crc kubenswrapper[4824]: I1209 11:23:58.519425 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5ace41b-1dfc-4f99-9ea8-c8ed32963685-catalog-content\") pod \"community-operators-n7r8x\" (UID: \"e5ace41b-1dfc-4f99-9ea8-c8ed32963685\") " pod="openshift-marketplace/community-operators-n7r8x" Dec 09 11:23:58 crc kubenswrapper[4824]: I1209 11:23:58.519674 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5ace41b-1dfc-4f99-9ea8-c8ed32963685-utilities\") pod \"community-operators-n7r8x\" (UID: \"e5ace41b-1dfc-4f99-9ea8-c8ed32963685\") " pod="openshift-marketplace/community-operators-n7r8x" Dec 09 11:23:58 crc kubenswrapper[4824]: I1209 11:23:58.519774 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkhfv\" (UniqueName: \"kubernetes.io/projected/e5ace41b-1dfc-4f99-9ea8-c8ed32963685-kube-api-access-lkhfv\") pod \"community-operators-n7r8x\" (UID: \"e5ace41b-1dfc-4f99-9ea8-c8ed32963685\") " pod="openshift-marketplace/community-operators-n7r8x" Dec 09 11:23:58 crc kubenswrapper[4824]: I1209 11:23:58.621616 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkhfv\" (UniqueName: \"kubernetes.io/projected/e5ace41b-1dfc-4f99-9ea8-c8ed32963685-kube-api-access-lkhfv\") pod \"community-operators-n7r8x\" (UID: \"e5ace41b-1dfc-4f99-9ea8-c8ed32963685\") " pod="openshift-marketplace/community-operators-n7r8x" Dec 09 11:23:58 crc kubenswrapper[4824]: I1209 11:23:58.622039 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5ace41b-1dfc-4f99-9ea8-c8ed32963685-catalog-content\") pod \"community-operators-n7r8x\" (UID: \"e5ace41b-1dfc-4f99-9ea8-c8ed32963685\") " pod="openshift-marketplace/community-operators-n7r8x" Dec 09 11:23:58 crc kubenswrapper[4824]: I1209 11:23:58.622218 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5ace41b-1dfc-4f99-9ea8-c8ed32963685-utilities\") pod \"community-operators-n7r8x\" (UID: \"e5ace41b-1dfc-4f99-9ea8-c8ed32963685\") " pod="openshift-marketplace/community-operators-n7r8x" Dec 09 11:23:58 crc kubenswrapper[4824]: I1209 11:23:58.622756 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5ace41b-1dfc-4f99-9ea8-c8ed32963685-catalog-content\") pod \"community-operators-n7r8x\" (UID: \"e5ace41b-1dfc-4f99-9ea8-c8ed32963685\") " pod="openshift-marketplace/community-operators-n7r8x" Dec 09 11:23:58 crc kubenswrapper[4824]: I1209 11:23:58.622793 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5ace41b-1dfc-4f99-9ea8-c8ed32963685-utilities\") pod \"community-operators-n7r8x\" (UID: \"e5ace41b-1dfc-4f99-9ea8-c8ed32963685\") " pod="openshift-marketplace/community-operators-n7r8x" Dec 09 11:23:58 crc kubenswrapper[4824]: I1209 11:23:58.657983 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkhfv\" (UniqueName: \"kubernetes.io/projected/e5ace41b-1dfc-4f99-9ea8-c8ed32963685-kube-api-access-lkhfv\") pod \"community-operators-n7r8x\" (UID: \"e5ace41b-1dfc-4f99-9ea8-c8ed32963685\") " pod="openshift-marketplace/community-operators-n7r8x" Dec 09 11:23:58 crc kubenswrapper[4824]: I1209 11:23:58.694753 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n7r8x" Dec 09 11:23:58 crc kubenswrapper[4824]: I1209 11:23:58.961759 4824 generic.go:334] "Generic (PLEG): container finished" podID="564a4ea7-ea14-41ee-be38-4eb8c5a22446" containerID="749218e46e0eebdfa29fdae704e00418fac0a74a14d921129a43c057d12f6c1d" exitCode=0 Dec 09 11:23:58 crc kubenswrapper[4824]: I1209 11:23:58.962078 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8fpgn" event={"ID":"564a4ea7-ea14-41ee-be38-4eb8c5a22446","Type":"ContainerDied","Data":"749218e46e0eebdfa29fdae704e00418fac0a74a14d921129a43c057d12f6c1d"} Dec 09 11:23:59 crc kubenswrapper[4824]: I1209 11:23:59.239626 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n7r8x"] Dec 09 11:23:59 crc kubenswrapper[4824]: I1209 11:23:59.981609 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7r8x" event={"ID":"e5ace41b-1dfc-4f99-9ea8-c8ed32963685","Type":"ContainerStarted","Data":"b5330f323d69256f476a85e49e78debbcf8bb8ab53977b382a96d2fd75cdb68d"} Dec 09 11:24:00 crc kubenswrapper[4824]: I1209 11:24:00.994098 4824 generic.go:334] "Generic (PLEG): container finished" podID="e5ace41b-1dfc-4f99-9ea8-c8ed32963685" containerID="58765548dc2d887ef2d0889176d836a9a7bcb02c79c825f6913d3768a7ac3075" exitCode=0 Dec 09 11:24:00 crc kubenswrapper[4824]: I1209 11:24:00.994178 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7r8x" event={"ID":"e5ace41b-1dfc-4f99-9ea8-c8ed32963685","Type":"ContainerDied","Data":"58765548dc2d887ef2d0889176d836a9a7bcb02c79c825f6913d3768a7ac3075"} Dec 09 11:24:00 crc kubenswrapper[4824]: I1209 11:24:00.998802 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8fpgn" event={"ID":"564a4ea7-ea14-41ee-be38-4eb8c5a22446","Type":"ContainerStarted","Data":"d108abd9c4ae47ff80a7be3f5048eec8e42e956d481b9b6ab88fb3623479280f"} Dec 09 11:24:01 crc kubenswrapper[4824]: I1209 11:24:01.049910 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8fpgn" podStartSLOduration=3.713182222 podStartE2EDuration="7.049892373s" podCreationTimestamp="2025-12-09 11:23:54 +0000 UTC" firstStartedPulling="2025-12-09 11:23:56.926342348 +0000 UTC m=+5793.260847025" lastFinishedPulling="2025-12-09 11:24:00.263052509 +0000 UTC m=+5796.597557176" observedRunningTime="2025-12-09 11:24:01.043322957 +0000 UTC m=+5797.377827624" watchObservedRunningTime="2025-12-09 11:24:01.049892373 +0000 UTC m=+5797.384397040" Dec 09 11:24:01 crc kubenswrapper[4824]: I1209 11:24:01.911259 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:24:01 crc kubenswrapper[4824]: E1209 11:24:01.913510 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:24:03 crc kubenswrapper[4824]: I1209 11:24:03.865818 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:24:03 crc kubenswrapper[4824]: I1209 11:24:03.866230 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:24:04 crc kubenswrapper[4824]: I1209 11:24:04.885478 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7r8x" event={"ID":"e5ace41b-1dfc-4f99-9ea8-c8ed32963685","Type":"ContainerStarted","Data":"4409a12530ab25e126ef46b02b4776306ec66237951b125b57323c8196871d24"} Dec 09 11:24:05 crc kubenswrapper[4824]: I1209 11:24:05.382606 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8fpgn" Dec 09 11:24:05 crc kubenswrapper[4824]: I1209 11:24:05.385963 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8fpgn" Dec 09 11:24:05 crc kubenswrapper[4824]: I1209 11:24:05.926730 4824 generic.go:334] "Generic (PLEG): container finished" podID="e5ace41b-1dfc-4f99-9ea8-c8ed32963685" containerID="4409a12530ab25e126ef46b02b4776306ec66237951b125b57323c8196871d24" exitCode=0 Dec 09 11:24:05 crc kubenswrapper[4824]: I1209 11:24:05.948353 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7r8x" event={"ID":"e5ace41b-1dfc-4f99-9ea8-c8ed32963685","Type":"ContainerDied","Data":"4409a12530ab25e126ef46b02b4776306ec66237951b125b57323c8196871d24"} Dec 09 11:24:06 crc kubenswrapper[4824]: I1209 11:24:06.477338 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-8fpgn" podUID="564a4ea7-ea14-41ee-be38-4eb8c5a22446" containerName="registry-server" probeResult="failure" output=< Dec 09 11:24:06 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 11:24:06 crc kubenswrapper[4824]: > Dec 09 11:24:06 crc kubenswrapper[4824]: I1209 11:24:06.942332 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7r8x" event={"ID":"e5ace41b-1dfc-4f99-9ea8-c8ed32963685","Type":"ContainerStarted","Data":"1883dd8ffc38bee065e2df22426da272eb2418e6ee09cade88c4f9cb2851f164"} Dec 09 11:24:06 crc kubenswrapper[4824]: I1209 11:24:06.965966 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n7r8x" podStartSLOduration=3.641467729 podStartE2EDuration="8.965942204s" podCreationTimestamp="2025-12-09 11:23:58 +0000 UTC" firstStartedPulling="2025-12-09 11:24:00.997231613 +0000 UTC m=+5797.331736280" lastFinishedPulling="2025-12-09 11:24:06.321706048 +0000 UTC m=+5802.656210755" observedRunningTime="2025-12-09 11:24:06.963057434 +0000 UTC m=+5803.297562111" watchObservedRunningTime="2025-12-09 11:24:06.965942204 +0000 UTC m=+5803.300446871" Dec 09 11:24:08 crc kubenswrapper[4824]: I1209 11:24:08.695853 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n7r8x" Dec 09 11:24:08 crc kubenswrapper[4824]: I1209 11:24:08.696108 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n7r8x" Dec 09 11:24:09 crc kubenswrapper[4824]: I1209 11:24:09.759925 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-n7r8x" podUID="e5ace41b-1dfc-4f99-9ea8-c8ed32963685" containerName="registry-server" probeResult="failure" output=< Dec 09 11:24:09 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 11:24:09 crc kubenswrapper[4824]: > Dec 09 11:24:11 crc kubenswrapper[4824]: I1209 11:24:11.601868 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-slv6d"] Dec 09 11:24:11 crc kubenswrapper[4824]: I1209 11:24:11.605680 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-slv6d" Dec 09 11:24:11 crc kubenswrapper[4824]: I1209 11:24:11.628753 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-slv6d"] Dec 09 11:24:11 crc kubenswrapper[4824]: I1209 11:24:11.786169 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8382995-7de6-49b8-acee-e797fd5d321c-utilities\") pod \"redhat-operators-slv6d\" (UID: \"a8382995-7de6-49b8-acee-e797fd5d321c\") " pod="openshift-marketplace/redhat-operators-slv6d" Dec 09 11:24:11 crc kubenswrapper[4824]: I1209 11:24:11.786347 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8382995-7de6-49b8-acee-e797fd5d321c-catalog-content\") pod \"redhat-operators-slv6d\" (UID: \"a8382995-7de6-49b8-acee-e797fd5d321c\") " pod="openshift-marketplace/redhat-operators-slv6d" Dec 09 11:24:11 crc kubenswrapper[4824]: I1209 11:24:11.786466 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfwl5\" (UniqueName: \"kubernetes.io/projected/a8382995-7de6-49b8-acee-e797fd5d321c-kube-api-access-bfwl5\") pod \"redhat-operators-slv6d\" (UID: \"a8382995-7de6-49b8-acee-e797fd5d321c\") " pod="openshift-marketplace/redhat-operators-slv6d" Dec 09 11:24:11 crc kubenswrapper[4824]: I1209 11:24:11.889346 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8382995-7de6-49b8-acee-e797fd5d321c-catalog-content\") pod \"redhat-operators-slv6d\" (UID: \"a8382995-7de6-49b8-acee-e797fd5d321c\") " pod="openshift-marketplace/redhat-operators-slv6d" Dec 09 11:24:11 crc kubenswrapper[4824]: I1209 11:24:11.889565 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfwl5\" (UniqueName: \"kubernetes.io/projected/a8382995-7de6-49b8-acee-e797fd5d321c-kube-api-access-bfwl5\") pod \"redhat-operators-slv6d\" (UID: \"a8382995-7de6-49b8-acee-e797fd5d321c\") " pod="openshift-marketplace/redhat-operators-slv6d" Dec 09 11:24:11 crc kubenswrapper[4824]: I1209 11:24:11.889730 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8382995-7de6-49b8-acee-e797fd5d321c-utilities\") pod \"redhat-operators-slv6d\" (UID: \"a8382995-7de6-49b8-acee-e797fd5d321c\") " pod="openshift-marketplace/redhat-operators-slv6d" Dec 09 11:24:11 crc kubenswrapper[4824]: I1209 11:24:11.889865 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8382995-7de6-49b8-acee-e797fd5d321c-catalog-content\") pod \"redhat-operators-slv6d\" (UID: \"a8382995-7de6-49b8-acee-e797fd5d321c\") " pod="openshift-marketplace/redhat-operators-slv6d" Dec 09 11:24:11 crc kubenswrapper[4824]: I1209 11:24:11.890452 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8382995-7de6-49b8-acee-e797fd5d321c-utilities\") pod \"redhat-operators-slv6d\" (UID: \"a8382995-7de6-49b8-acee-e797fd5d321c\") " pod="openshift-marketplace/redhat-operators-slv6d" Dec 09 11:24:11 crc kubenswrapper[4824]: I1209 11:24:11.928284 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfwl5\" (UniqueName: \"kubernetes.io/projected/a8382995-7de6-49b8-acee-e797fd5d321c-kube-api-access-bfwl5\") pod \"redhat-operators-slv6d\" (UID: \"a8382995-7de6-49b8-acee-e797fd5d321c\") " pod="openshift-marketplace/redhat-operators-slv6d" Dec 09 11:24:11 crc kubenswrapper[4824]: I1209 11:24:11.930822 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-slv6d" Dec 09 11:24:12 crc kubenswrapper[4824]: I1209 11:24:12.579070 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-slv6d"] Dec 09 11:24:13 crc kubenswrapper[4824]: I1209 11:24:13.017940 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-slv6d" event={"ID":"a8382995-7de6-49b8-acee-e797fd5d321c","Type":"ContainerStarted","Data":"d178f1e05624622f1cc82cba82452aa4ebdd059d81da042a43255780421987ab"} Dec 09 11:24:13 crc kubenswrapper[4824]: I1209 11:24:13.929829 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:24:13 crc kubenswrapper[4824]: E1209 11:24:13.932681 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:24:14 crc kubenswrapper[4824]: I1209 11:24:14.032453 4824 generic.go:334] "Generic (PLEG): container finished" podID="a8382995-7de6-49b8-acee-e797fd5d321c" containerID="f7f602e827fe94cbcb01d2ce598c2c013e1271f1795381a9931edfeb605bd2ef" exitCode=0 Dec 09 11:24:14 crc kubenswrapper[4824]: I1209 11:24:14.032515 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-slv6d" event={"ID":"a8382995-7de6-49b8-acee-e797fd5d321c","Type":"ContainerDied","Data":"f7f602e827fe94cbcb01d2ce598c2c013e1271f1795381a9931edfeb605bd2ef"} Dec 09 11:24:16 crc kubenswrapper[4824]: I1209 11:24:16.055777 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-slv6d" event={"ID":"a8382995-7de6-49b8-acee-e797fd5d321c","Type":"ContainerStarted","Data":"f90c6378c94ea4f7d327901a9647c7ea29153c03b0495ea10df5cf25b867f804"} Dec 09 11:24:16 crc kubenswrapper[4824]: I1209 11:24:16.222449 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8fpgn" Dec 09 11:24:16 crc kubenswrapper[4824]: I1209 11:24:16.297370 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8fpgn" Dec 09 11:24:17 crc kubenswrapper[4824]: I1209 11:24:17.979029 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8fpgn"] Dec 09 11:24:18 crc kubenswrapper[4824]: I1209 11:24:18.075590 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8fpgn" podUID="564a4ea7-ea14-41ee-be38-4eb8c5a22446" containerName="registry-server" containerID="cri-o://d108abd9c4ae47ff80a7be3f5048eec8e42e956d481b9b6ab88fb3623479280f" gracePeriod=2 Dec 09 11:24:18 crc kubenswrapper[4824]: I1209 11:24:18.761121 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n7r8x" Dec 09 11:24:18 crc kubenswrapper[4824]: I1209 11:24:18.825123 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n7r8x" Dec 09 11:24:19 crc kubenswrapper[4824]: I1209 11:24:19.090085 4824 generic.go:334] "Generic (PLEG): container finished" podID="564a4ea7-ea14-41ee-be38-4eb8c5a22446" containerID="d108abd9c4ae47ff80a7be3f5048eec8e42e956d481b9b6ab88fb3623479280f" exitCode=0 Dec 09 11:24:19 crc kubenswrapper[4824]: I1209 11:24:19.090312 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8fpgn" event={"ID":"564a4ea7-ea14-41ee-be38-4eb8c5a22446","Type":"ContainerDied","Data":"d108abd9c4ae47ff80a7be3f5048eec8e42e956d481b9b6ab88fb3623479280f"} Dec 09 11:24:19 crc kubenswrapper[4824]: I1209 11:24:19.383184 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8fpgn" Dec 09 11:24:19 crc kubenswrapper[4824]: I1209 11:24:19.441628 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/564a4ea7-ea14-41ee-be38-4eb8c5a22446-utilities\") pod \"564a4ea7-ea14-41ee-be38-4eb8c5a22446\" (UID: \"564a4ea7-ea14-41ee-be38-4eb8c5a22446\") " Dec 09 11:24:19 crc kubenswrapper[4824]: I1209 11:24:19.441759 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rh4d8\" (UniqueName: \"kubernetes.io/projected/564a4ea7-ea14-41ee-be38-4eb8c5a22446-kube-api-access-rh4d8\") pod \"564a4ea7-ea14-41ee-be38-4eb8c5a22446\" (UID: \"564a4ea7-ea14-41ee-be38-4eb8c5a22446\") " Dec 09 11:24:19 crc kubenswrapper[4824]: I1209 11:24:19.441970 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/564a4ea7-ea14-41ee-be38-4eb8c5a22446-catalog-content\") pod \"564a4ea7-ea14-41ee-be38-4eb8c5a22446\" (UID: \"564a4ea7-ea14-41ee-be38-4eb8c5a22446\") " Dec 09 11:24:19 crc kubenswrapper[4824]: I1209 11:24:19.488059 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/564a4ea7-ea14-41ee-be38-4eb8c5a22446-utilities" (OuterVolumeSpecName: "utilities") pod "564a4ea7-ea14-41ee-be38-4eb8c5a22446" (UID: "564a4ea7-ea14-41ee-be38-4eb8c5a22446"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:24:19 crc kubenswrapper[4824]: I1209 11:24:19.499763 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/564a4ea7-ea14-41ee-be38-4eb8c5a22446-kube-api-access-rh4d8" (OuterVolumeSpecName: "kube-api-access-rh4d8") pod "564a4ea7-ea14-41ee-be38-4eb8c5a22446" (UID: "564a4ea7-ea14-41ee-be38-4eb8c5a22446"). InnerVolumeSpecName "kube-api-access-rh4d8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:24:19 crc kubenswrapper[4824]: I1209 11:24:19.545729 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/564a4ea7-ea14-41ee-be38-4eb8c5a22446-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:24:19 crc kubenswrapper[4824]: I1209 11:24:19.545774 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rh4d8\" (UniqueName: \"kubernetes.io/projected/564a4ea7-ea14-41ee-be38-4eb8c5a22446-kube-api-access-rh4d8\") on node \"crc\" DevicePath \"\"" Dec 09 11:24:19 crc kubenswrapper[4824]: I1209 11:24:19.552896 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/564a4ea7-ea14-41ee-be38-4eb8c5a22446-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "564a4ea7-ea14-41ee-be38-4eb8c5a22446" (UID: "564a4ea7-ea14-41ee-be38-4eb8c5a22446"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:24:19 crc kubenswrapper[4824]: I1209 11:24:19.648575 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/564a4ea7-ea14-41ee-be38-4eb8c5a22446-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:24:20 crc kubenswrapper[4824]: I1209 11:24:20.106214 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8fpgn" event={"ID":"564a4ea7-ea14-41ee-be38-4eb8c5a22446","Type":"ContainerDied","Data":"341efcfdc037b7c277571b71ed078aae2c032062fafa9e38eb5b8477707e2f99"} Dec 09 11:24:20 crc kubenswrapper[4824]: I1209 11:24:20.106268 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8fpgn" Dec 09 11:24:20 crc kubenswrapper[4824]: I1209 11:24:20.106287 4824 scope.go:117] "RemoveContainer" containerID="d108abd9c4ae47ff80a7be3f5048eec8e42e956d481b9b6ab88fb3623479280f" Dec 09 11:24:20 crc kubenswrapper[4824]: I1209 11:24:20.113409 4824 generic.go:334] "Generic (PLEG): container finished" podID="a8382995-7de6-49b8-acee-e797fd5d321c" containerID="f90c6378c94ea4f7d327901a9647c7ea29153c03b0495ea10df5cf25b867f804" exitCode=0 Dec 09 11:24:20 crc kubenswrapper[4824]: I1209 11:24:20.113449 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-slv6d" event={"ID":"a8382995-7de6-49b8-acee-e797fd5d321c","Type":"ContainerDied","Data":"f90c6378c94ea4f7d327901a9647c7ea29153c03b0495ea10df5cf25b867f804"} Dec 09 11:24:20 crc kubenswrapper[4824]: I1209 11:24:20.156754 4824 scope.go:117] "RemoveContainer" containerID="749218e46e0eebdfa29fdae704e00418fac0a74a14d921129a43c057d12f6c1d" Dec 09 11:24:20 crc kubenswrapper[4824]: I1209 11:24:20.156834 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8fpgn"] Dec 09 11:24:20 crc kubenswrapper[4824]: I1209 11:24:20.195442 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8fpgn"] Dec 09 11:24:20 crc kubenswrapper[4824]: I1209 11:24:20.195747 4824 scope.go:117] "RemoveContainer" containerID="c9cd33cc3af706051ab8073731d7acb29b927e424a2f81990f7a70ff4bf6f5db" Dec 09 11:24:21 crc kubenswrapper[4824]: I1209 11:24:21.141796 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-slv6d" event={"ID":"a8382995-7de6-49b8-acee-e797fd5d321c","Type":"ContainerStarted","Data":"07a7d6dc1a70e071b087fdb2f738bc00adaa50fecdc9c4ae275e604b65bf4604"} Dec 09 11:24:21 crc kubenswrapper[4824]: I1209 11:24:21.177825 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n7r8x"] Dec 09 11:24:21 crc kubenswrapper[4824]: I1209 11:24:21.178109 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n7r8x" podUID="e5ace41b-1dfc-4f99-9ea8-c8ed32963685" containerName="registry-server" containerID="cri-o://1883dd8ffc38bee065e2df22426da272eb2418e6ee09cade88c4f9cb2851f164" gracePeriod=2 Dec 09 11:24:21 crc kubenswrapper[4824]: I1209 11:24:21.185656 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-slv6d" podStartSLOduration=3.613505509 podStartE2EDuration="10.185624577s" podCreationTimestamp="2025-12-09 11:24:11 +0000 UTC" firstStartedPulling="2025-12-09 11:24:14.034567319 +0000 UTC m=+5810.369071986" lastFinishedPulling="2025-12-09 11:24:20.606686387 +0000 UTC m=+5816.941191054" observedRunningTime="2025-12-09 11:24:21.178565896 +0000 UTC m=+5817.513070583" watchObservedRunningTime="2025-12-09 11:24:21.185624577 +0000 UTC m=+5817.520129244" Dec 09 11:24:21 crc kubenswrapper[4824]: I1209 11:24:21.778734 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n7r8x" Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.072523 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-slv6d" Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.076064 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5ace41b-1dfc-4f99-9ea8-c8ed32963685-utilities\") pod \"e5ace41b-1dfc-4f99-9ea8-c8ed32963685\" (UID: \"e5ace41b-1dfc-4f99-9ea8-c8ed32963685\") " Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.076245 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkhfv\" (UniqueName: \"kubernetes.io/projected/e5ace41b-1dfc-4f99-9ea8-c8ed32963685-kube-api-access-lkhfv\") pod \"e5ace41b-1dfc-4f99-9ea8-c8ed32963685\" (UID: \"e5ace41b-1dfc-4f99-9ea8-c8ed32963685\") " Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.076498 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5ace41b-1dfc-4f99-9ea8-c8ed32963685-catalog-content\") pod \"e5ace41b-1dfc-4f99-9ea8-c8ed32963685\" (UID: \"e5ace41b-1dfc-4f99-9ea8-c8ed32963685\") " Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.079677 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5ace41b-1dfc-4f99-9ea8-c8ed32963685-utilities" (OuterVolumeSpecName: "utilities") pod "e5ace41b-1dfc-4f99-9ea8-c8ed32963685" (UID: "e5ace41b-1dfc-4f99-9ea8-c8ed32963685"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.091615 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5ace41b-1dfc-4f99-9ea8-c8ed32963685-kube-api-access-lkhfv" (OuterVolumeSpecName: "kube-api-access-lkhfv") pod "e5ace41b-1dfc-4f99-9ea8-c8ed32963685" (UID: "e5ace41b-1dfc-4f99-9ea8-c8ed32963685"). InnerVolumeSpecName "kube-api-access-lkhfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.119339 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="564a4ea7-ea14-41ee-be38-4eb8c5a22446" path="/var/lib/kubelet/pods/564a4ea7-ea14-41ee-be38-4eb8c5a22446/volumes" Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.144412 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-slv6d" Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.162265 4824 generic.go:334] "Generic (PLEG): container finished" podID="e5ace41b-1dfc-4f99-9ea8-c8ed32963685" containerID="1883dd8ffc38bee065e2df22426da272eb2418e6ee09cade88c4f9cb2851f164" exitCode=0 Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.162382 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n7r8x" Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.162375 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7r8x" event={"ID":"e5ace41b-1dfc-4f99-9ea8-c8ed32963685","Type":"ContainerDied","Data":"1883dd8ffc38bee065e2df22426da272eb2418e6ee09cade88c4f9cb2851f164"} Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.162652 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7r8x" event={"ID":"e5ace41b-1dfc-4f99-9ea8-c8ed32963685","Type":"ContainerDied","Data":"b5330f323d69256f476a85e49e78debbcf8bb8ab53977b382a96d2fd75cdb68d"} Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.162684 4824 scope.go:117] "RemoveContainer" containerID="1883dd8ffc38bee065e2df22426da272eb2418e6ee09cade88c4f9cb2851f164" Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.179837 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5ace41b-1dfc-4f99-9ea8-c8ed32963685-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5ace41b-1dfc-4f99-9ea8-c8ed32963685" (UID: "e5ace41b-1dfc-4f99-9ea8-c8ed32963685"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.183068 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5ace41b-1dfc-4f99-9ea8-c8ed32963685-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.183202 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkhfv\" (UniqueName: \"kubernetes.io/projected/e5ace41b-1dfc-4f99-9ea8-c8ed32963685-kube-api-access-lkhfv\") on node \"crc\" DevicePath \"\"" Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.183274 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5ace41b-1dfc-4f99-9ea8-c8ed32963685-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.187686 4824 scope.go:117] "RemoveContainer" containerID="4409a12530ab25e126ef46b02b4776306ec66237951b125b57323c8196871d24" Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.217195 4824 scope.go:117] "RemoveContainer" containerID="58765548dc2d887ef2d0889176d836a9a7bcb02c79c825f6913d3768a7ac3075" Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.267212 4824 scope.go:117] "RemoveContainer" containerID="1883dd8ffc38bee065e2df22426da272eb2418e6ee09cade88c4f9cb2851f164" Dec 09 11:24:22 crc kubenswrapper[4824]: E1209 11:24:22.267836 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1883dd8ffc38bee065e2df22426da272eb2418e6ee09cade88c4f9cb2851f164\": container with ID starting with 1883dd8ffc38bee065e2df22426da272eb2418e6ee09cade88c4f9cb2851f164 not found: ID does not exist" containerID="1883dd8ffc38bee065e2df22426da272eb2418e6ee09cade88c4f9cb2851f164" Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.268332 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1883dd8ffc38bee065e2df22426da272eb2418e6ee09cade88c4f9cb2851f164"} err="failed to get container status \"1883dd8ffc38bee065e2df22426da272eb2418e6ee09cade88c4f9cb2851f164\": rpc error: code = NotFound desc = could not find container \"1883dd8ffc38bee065e2df22426da272eb2418e6ee09cade88c4f9cb2851f164\": container with ID starting with 1883dd8ffc38bee065e2df22426da272eb2418e6ee09cade88c4f9cb2851f164 not found: ID does not exist" Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.268435 4824 scope.go:117] "RemoveContainer" containerID="4409a12530ab25e126ef46b02b4776306ec66237951b125b57323c8196871d24" Dec 09 11:24:22 crc kubenswrapper[4824]: E1209 11:24:22.268742 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4409a12530ab25e126ef46b02b4776306ec66237951b125b57323c8196871d24\": container with ID starting with 4409a12530ab25e126ef46b02b4776306ec66237951b125b57323c8196871d24 not found: ID does not exist" containerID="4409a12530ab25e126ef46b02b4776306ec66237951b125b57323c8196871d24" Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.268891 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4409a12530ab25e126ef46b02b4776306ec66237951b125b57323c8196871d24"} err="failed to get container status \"4409a12530ab25e126ef46b02b4776306ec66237951b125b57323c8196871d24\": rpc error: code = NotFound desc = could not find container \"4409a12530ab25e126ef46b02b4776306ec66237951b125b57323c8196871d24\": container with ID starting with 4409a12530ab25e126ef46b02b4776306ec66237951b125b57323c8196871d24 not found: ID does not exist" Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.268973 4824 scope.go:117] "RemoveContainer" containerID="58765548dc2d887ef2d0889176d836a9a7bcb02c79c825f6913d3768a7ac3075" Dec 09 11:24:22 crc kubenswrapper[4824]: E1209 11:24:22.269283 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58765548dc2d887ef2d0889176d836a9a7bcb02c79c825f6913d3768a7ac3075\": container with ID starting with 58765548dc2d887ef2d0889176d836a9a7bcb02c79c825f6913d3768a7ac3075 not found: ID does not exist" containerID="58765548dc2d887ef2d0889176d836a9a7bcb02c79c825f6913d3768a7ac3075" Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.269388 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58765548dc2d887ef2d0889176d836a9a7bcb02c79c825f6913d3768a7ac3075"} err="failed to get container status \"58765548dc2d887ef2d0889176d836a9a7bcb02c79c825f6913d3768a7ac3075\": rpc error: code = NotFound desc = could not find container \"58765548dc2d887ef2d0889176d836a9a7bcb02c79c825f6913d3768a7ac3075\": container with ID starting with 58765548dc2d887ef2d0889176d836a9a7bcb02c79c825f6913d3768a7ac3075 not found: ID does not exist" Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.506763 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n7r8x"] Dec 09 11:24:22 crc kubenswrapper[4824]: I1209 11:24:22.523283 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n7r8x"] Dec 09 11:24:23 crc kubenswrapper[4824]: I1209 11:24:23.152218 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-slv6d" podUID="a8382995-7de6-49b8-acee-e797fd5d321c" containerName="registry-server" probeResult="failure" output=< Dec 09 11:24:23 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 11:24:23 crc kubenswrapper[4824]: > Dec 09 11:24:23 crc kubenswrapper[4824]: I1209 11:24:23.990351 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5ace41b-1dfc-4f99-9ea8-c8ed32963685" path="/var/lib/kubelet/pods/e5ace41b-1dfc-4f99-9ea8-c8ed32963685/volumes" Dec 09 11:24:28 crc kubenswrapper[4824]: I1209 11:24:28.911175 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:24:28 crc kubenswrapper[4824]: E1209 11:24:28.915361 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:24:32 crc kubenswrapper[4824]: I1209 11:24:32.861216 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:24:32 crc kubenswrapper[4824]: I1209 11:24:32.861739 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:24:32 crc kubenswrapper[4824]: I1209 11:24:32.861813 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 11:24:32 crc kubenswrapper[4824]: I1209 11:24:32.862692 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"49a74dd2910658c46255de301a11501ac6eaebf752b480f0958816296c35a0f8"} pod="openshift-machine-config-operator/machine-config-daemon-dth8x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 11:24:32 crc kubenswrapper[4824]: I1209 11:24:32.862742 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" containerID="cri-o://49a74dd2910658c46255de301a11501ac6eaebf752b480f0958816296c35a0f8" gracePeriod=600 Dec 09 11:24:33 crc kubenswrapper[4824]: I1209 11:24:33.023418 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-slv6d" podUID="a8382995-7de6-49b8-acee-e797fd5d321c" containerName="registry-server" probeResult="failure" output=< Dec 09 11:24:33 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Dec 09 11:24:33 crc kubenswrapper[4824]: > Dec 09 11:24:33 crc kubenswrapper[4824]: I1209 11:24:33.457243 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerID="49a74dd2910658c46255de301a11501ac6eaebf752b480f0958816296c35a0f8" exitCode=0 Dec 09 11:24:33 crc kubenswrapper[4824]: I1209 11:24:33.457340 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerDied","Data":"49a74dd2910658c46255de301a11501ac6eaebf752b480f0958816296c35a0f8"} Dec 09 11:24:33 crc kubenswrapper[4824]: I1209 11:24:33.457643 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerStarted","Data":"34c807c5b217ed0d4c39e0fefa50a143310147aedb732f65dad4bdb215fe26a8"} Dec 09 11:24:33 crc kubenswrapper[4824]: I1209 11:24:33.457672 4824 scope.go:117] "RemoveContainer" containerID="84c411c148790b75818c0bc90f3916f7d511650011c9cc5b9ab38e4d988c0d20" Dec 09 11:24:41 crc kubenswrapper[4824]: I1209 11:24:41.912621 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:24:41 crc kubenswrapper[4824]: E1209 11:24:41.916305 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:24:42 crc kubenswrapper[4824]: I1209 11:24:42.008219 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-slv6d" Dec 09 11:24:42 crc kubenswrapper[4824]: I1209 11:24:42.070291 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-slv6d" Dec 09 11:24:42 crc kubenswrapper[4824]: I1209 11:24:42.802163 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-slv6d"] Dec 09 11:24:43 crc kubenswrapper[4824]: I1209 11:24:43.600947 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-slv6d" podUID="a8382995-7de6-49b8-acee-e797fd5d321c" containerName="registry-server" containerID="cri-o://07a7d6dc1a70e071b087fdb2f738bc00adaa50fecdc9c4ae275e604b65bf4604" gracePeriod=2 Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.182825 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-slv6d" Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.319279 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfwl5\" (UniqueName: \"kubernetes.io/projected/a8382995-7de6-49b8-acee-e797fd5d321c-kube-api-access-bfwl5\") pod \"a8382995-7de6-49b8-acee-e797fd5d321c\" (UID: \"a8382995-7de6-49b8-acee-e797fd5d321c\") " Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.319498 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8382995-7de6-49b8-acee-e797fd5d321c-catalog-content\") pod \"a8382995-7de6-49b8-acee-e797fd5d321c\" (UID: \"a8382995-7de6-49b8-acee-e797fd5d321c\") " Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.319710 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8382995-7de6-49b8-acee-e797fd5d321c-utilities\") pod \"a8382995-7de6-49b8-acee-e797fd5d321c\" (UID: \"a8382995-7de6-49b8-acee-e797fd5d321c\") " Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.320585 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8382995-7de6-49b8-acee-e797fd5d321c-utilities" (OuterVolumeSpecName: "utilities") pod "a8382995-7de6-49b8-acee-e797fd5d321c" (UID: "a8382995-7de6-49b8-acee-e797fd5d321c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.350178 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8382995-7de6-49b8-acee-e797fd5d321c-kube-api-access-bfwl5" (OuterVolumeSpecName: "kube-api-access-bfwl5") pod "a8382995-7de6-49b8-acee-e797fd5d321c" (UID: "a8382995-7de6-49b8-acee-e797fd5d321c"). InnerVolumeSpecName "kube-api-access-bfwl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.423309 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8382995-7de6-49b8-acee-e797fd5d321c-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.423594 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfwl5\" (UniqueName: \"kubernetes.io/projected/a8382995-7de6-49b8-acee-e797fd5d321c-kube-api-access-bfwl5\") on node \"crc\" DevicePath \"\"" Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.448214 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8382995-7de6-49b8-acee-e797fd5d321c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a8382995-7de6-49b8-acee-e797fd5d321c" (UID: "a8382995-7de6-49b8-acee-e797fd5d321c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.525948 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8382995-7de6-49b8-acee-e797fd5d321c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.617065 4824 generic.go:334] "Generic (PLEG): container finished" podID="a8382995-7de6-49b8-acee-e797fd5d321c" containerID="07a7d6dc1a70e071b087fdb2f738bc00adaa50fecdc9c4ae275e604b65bf4604" exitCode=0 Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.617114 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-slv6d" event={"ID":"a8382995-7de6-49b8-acee-e797fd5d321c","Type":"ContainerDied","Data":"07a7d6dc1a70e071b087fdb2f738bc00adaa50fecdc9c4ae275e604b65bf4604"} Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.617145 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-slv6d" event={"ID":"a8382995-7de6-49b8-acee-e797fd5d321c","Type":"ContainerDied","Data":"d178f1e05624622f1cc82cba82452aa4ebdd059d81da042a43255780421987ab"} Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.617164 4824 scope.go:117] "RemoveContainer" containerID="07a7d6dc1a70e071b087fdb2f738bc00adaa50fecdc9c4ae275e604b65bf4604" Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.617183 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-slv6d" Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.653807 4824 scope.go:117] "RemoveContainer" containerID="f90c6378c94ea4f7d327901a9647c7ea29153c03b0495ea10df5cf25b867f804" Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.661162 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-slv6d"] Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.678874 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-slv6d"] Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.683695 4824 scope.go:117] "RemoveContainer" containerID="f7f602e827fe94cbcb01d2ce598c2c013e1271f1795381a9931edfeb605bd2ef" Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.730196 4824 scope.go:117] "RemoveContainer" containerID="07a7d6dc1a70e071b087fdb2f738bc00adaa50fecdc9c4ae275e604b65bf4604" Dec 09 11:24:44 crc kubenswrapper[4824]: E1209 11:24:44.730684 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07a7d6dc1a70e071b087fdb2f738bc00adaa50fecdc9c4ae275e604b65bf4604\": container with ID starting with 07a7d6dc1a70e071b087fdb2f738bc00adaa50fecdc9c4ae275e604b65bf4604 not found: ID does not exist" containerID="07a7d6dc1a70e071b087fdb2f738bc00adaa50fecdc9c4ae275e604b65bf4604" Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.730748 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07a7d6dc1a70e071b087fdb2f738bc00adaa50fecdc9c4ae275e604b65bf4604"} err="failed to get container status \"07a7d6dc1a70e071b087fdb2f738bc00adaa50fecdc9c4ae275e604b65bf4604\": rpc error: code = NotFound desc = could not find container \"07a7d6dc1a70e071b087fdb2f738bc00adaa50fecdc9c4ae275e604b65bf4604\": container with ID starting with 07a7d6dc1a70e071b087fdb2f738bc00adaa50fecdc9c4ae275e604b65bf4604 not found: ID does not exist" Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.730805 4824 scope.go:117] "RemoveContainer" containerID="f90c6378c94ea4f7d327901a9647c7ea29153c03b0495ea10df5cf25b867f804" Dec 09 11:24:44 crc kubenswrapper[4824]: E1209 11:24:44.731422 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f90c6378c94ea4f7d327901a9647c7ea29153c03b0495ea10df5cf25b867f804\": container with ID starting with f90c6378c94ea4f7d327901a9647c7ea29153c03b0495ea10df5cf25b867f804 not found: ID does not exist" containerID="f90c6378c94ea4f7d327901a9647c7ea29153c03b0495ea10df5cf25b867f804" Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.731470 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f90c6378c94ea4f7d327901a9647c7ea29153c03b0495ea10df5cf25b867f804"} err="failed to get container status \"f90c6378c94ea4f7d327901a9647c7ea29153c03b0495ea10df5cf25b867f804\": rpc error: code = NotFound desc = could not find container \"f90c6378c94ea4f7d327901a9647c7ea29153c03b0495ea10df5cf25b867f804\": container with ID starting with f90c6378c94ea4f7d327901a9647c7ea29153c03b0495ea10df5cf25b867f804 not found: ID does not exist" Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.731499 4824 scope.go:117] "RemoveContainer" containerID="f7f602e827fe94cbcb01d2ce598c2c013e1271f1795381a9931edfeb605bd2ef" Dec 09 11:24:44 crc kubenswrapper[4824]: E1209 11:24:44.731805 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7f602e827fe94cbcb01d2ce598c2c013e1271f1795381a9931edfeb605bd2ef\": container with ID starting with f7f602e827fe94cbcb01d2ce598c2c013e1271f1795381a9931edfeb605bd2ef not found: ID does not exist" containerID="f7f602e827fe94cbcb01d2ce598c2c013e1271f1795381a9931edfeb605bd2ef" Dec 09 11:24:44 crc kubenswrapper[4824]: I1209 11:24:44.731837 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7f602e827fe94cbcb01d2ce598c2c013e1271f1795381a9931edfeb605bd2ef"} err="failed to get container status \"f7f602e827fe94cbcb01d2ce598c2c013e1271f1795381a9931edfeb605bd2ef\": rpc error: code = NotFound desc = could not find container \"f7f602e827fe94cbcb01d2ce598c2c013e1271f1795381a9931edfeb605bd2ef\": container with ID starting with f7f602e827fe94cbcb01d2ce598c2c013e1271f1795381a9931edfeb605bd2ef not found: ID does not exist" Dec 09 11:24:45 crc kubenswrapper[4824]: I1209 11:24:45.925473 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8382995-7de6-49b8-acee-e797fd5d321c" path="/var/lib/kubelet/pods/a8382995-7de6-49b8-acee-e797fd5d321c/volumes" Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.214624 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cw5vt"] Dec 09 11:24:47 crc kubenswrapper[4824]: E1209 11:24:47.215576 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="564a4ea7-ea14-41ee-be38-4eb8c5a22446" containerName="registry-server" Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.215594 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="564a4ea7-ea14-41ee-be38-4eb8c5a22446" containerName="registry-server" Dec 09 11:24:47 crc kubenswrapper[4824]: E1209 11:24:47.215693 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="564a4ea7-ea14-41ee-be38-4eb8c5a22446" containerName="extract-utilities" Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.215703 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="564a4ea7-ea14-41ee-be38-4eb8c5a22446" containerName="extract-utilities" Dec 09 11:24:47 crc kubenswrapper[4824]: E1209 11:24:47.215740 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8382995-7de6-49b8-acee-e797fd5d321c" containerName="extract-content" Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.215749 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8382995-7de6-49b8-acee-e797fd5d321c" containerName="extract-content" Dec 09 11:24:47 crc kubenswrapper[4824]: E1209 11:24:47.215808 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8382995-7de6-49b8-acee-e797fd5d321c" containerName="extract-utilities" Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.215818 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8382995-7de6-49b8-acee-e797fd5d321c" containerName="extract-utilities" Dec 09 11:24:47 crc kubenswrapper[4824]: E1209 11:24:47.215838 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5ace41b-1dfc-4f99-9ea8-c8ed32963685" containerName="extract-content" Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.215846 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5ace41b-1dfc-4f99-9ea8-c8ed32963685" containerName="extract-content" Dec 09 11:24:47 crc kubenswrapper[4824]: E1209 11:24:47.215873 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8382995-7de6-49b8-acee-e797fd5d321c" containerName="registry-server" Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.215881 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8382995-7de6-49b8-acee-e797fd5d321c" containerName="registry-server" Dec 09 11:24:47 crc kubenswrapper[4824]: E1209 11:24:47.215897 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="564a4ea7-ea14-41ee-be38-4eb8c5a22446" containerName="extract-content" Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.215906 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="564a4ea7-ea14-41ee-be38-4eb8c5a22446" containerName="extract-content" Dec 09 11:24:47 crc kubenswrapper[4824]: E1209 11:24:47.215944 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5ace41b-1dfc-4f99-9ea8-c8ed32963685" containerName="extract-utilities" Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.215953 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5ace41b-1dfc-4f99-9ea8-c8ed32963685" containerName="extract-utilities" Dec 09 11:24:47 crc kubenswrapper[4824]: E1209 11:24:47.215973 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5ace41b-1dfc-4f99-9ea8-c8ed32963685" containerName="registry-server" Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.215982 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5ace41b-1dfc-4f99-9ea8-c8ed32963685" containerName="registry-server" Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.216284 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="564a4ea7-ea14-41ee-be38-4eb8c5a22446" containerName="registry-server" Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.216328 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5ace41b-1dfc-4f99-9ea8-c8ed32963685" containerName="registry-server" Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.216353 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8382995-7de6-49b8-acee-e797fd5d321c" containerName="registry-server" Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.218807 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cw5vt" Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.221414 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9fnw\" (UniqueName: \"kubernetes.io/projected/7eab3325-7b88-47c2-b7cd-1a727a714b73-kube-api-access-z9fnw\") pod \"redhat-marketplace-cw5vt\" (UID: \"7eab3325-7b88-47c2-b7cd-1a727a714b73\") " pod="openshift-marketplace/redhat-marketplace-cw5vt" Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.221572 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eab3325-7b88-47c2-b7cd-1a727a714b73-catalog-content\") pod \"redhat-marketplace-cw5vt\" (UID: \"7eab3325-7b88-47c2-b7cd-1a727a714b73\") " pod="openshift-marketplace/redhat-marketplace-cw5vt" Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.221602 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eab3325-7b88-47c2-b7cd-1a727a714b73-utilities\") pod \"redhat-marketplace-cw5vt\" (UID: \"7eab3325-7b88-47c2-b7cd-1a727a714b73\") " pod="openshift-marketplace/redhat-marketplace-cw5vt" Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.227295 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cw5vt"] Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.324745 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9fnw\" (UniqueName: \"kubernetes.io/projected/7eab3325-7b88-47c2-b7cd-1a727a714b73-kube-api-access-z9fnw\") pod \"redhat-marketplace-cw5vt\" (UID: \"7eab3325-7b88-47c2-b7cd-1a727a714b73\") " pod="openshift-marketplace/redhat-marketplace-cw5vt" Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.324908 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eab3325-7b88-47c2-b7cd-1a727a714b73-catalog-content\") pod \"redhat-marketplace-cw5vt\" (UID: \"7eab3325-7b88-47c2-b7cd-1a727a714b73\") " pod="openshift-marketplace/redhat-marketplace-cw5vt" Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.324939 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eab3325-7b88-47c2-b7cd-1a727a714b73-utilities\") pod \"redhat-marketplace-cw5vt\" (UID: \"7eab3325-7b88-47c2-b7cd-1a727a714b73\") " pod="openshift-marketplace/redhat-marketplace-cw5vt" Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.325661 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eab3325-7b88-47c2-b7cd-1a727a714b73-catalog-content\") pod \"redhat-marketplace-cw5vt\" (UID: \"7eab3325-7b88-47c2-b7cd-1a727a714b73\") " pod="openshift-marketplace/redhat-marketplace-cw5vt" Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.325686 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eab3325-7b88-47c2-b7cd-1a727a714b73-utilities\") pod \"redhat-marketplace-cw5vt\" (UID: \"7eab3325-7b88-47c2-b7cd-1a727a714b73\") " pod="openshift-marketplace/redhat-marketplace-cw5vt" Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.346507 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9fnw\" (UniqueName: \"kubernetes.io/projected/7eab3325-7b88-47c2-b7cd-1a727a714b73-kube-api-access-z9fnw\") pod \"redhat-marketplace-cw5vt\" (UID: \"7eab3325-7b88-47c2-b7cd-1a727a714b73\") " pod="openshift-marketplace/redhat-marketplace-cw5vt" Dec 09 11:24:47 crc kubenswrapper[4824]: I1209 11:24:47.545387 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cw5vt" Dec 09 11:24:48 crc kubenswrapper[4824]: I1209 11:24:48.067116 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cw5vt"] Dec 09 11:24:48 crc kubenswrapper[4824]: W1209 11:24:48.074928 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7eab3325_7b88_47c2_b7cd_1a727a714b73.slice/crio-8a9edc7a12b4d4811920527ade45b7ba6e9a8a58fb2ea443f30de7b907e61edb WatchSource:0}: Error finding container 8a9edc7a12b4d4811920527ade45b7ba6e9a8a58fb2ea443f30de7b907e61edb: Status 404 returned error can't find the container with id 8a9edc7a12b4d4811920527ade45b7ba6e9a8a58fb2ea443f30de7b907e61edb Dec 09 11:24:48 crc kubenswrapper[4824]: I1209 11:24:48.695375 4824 generic.go:334] "Generic (PLEG): container finished" podID="7eab3325-7b88-47c2-b7cd-1a727a714b73" containerID="493cc0a6fc4caf872a8d56dfd435fe2b27a8c288e308a6bac0a519b9b9674eab" exitCode=0 Dec 09 11:24:48 crc kubenswrapper[4824]: I1209 11:24:48.695687 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cw5vt" event={"ID":"7eab3325-7b88-47c2-b7cd-1a727a714b73","Type":"ContainerDied","Data":"493cc0a6fc4caf872a8d56dfd435fe2b27a8c288e308a6bac0a519b9b9674eab"} Dec 09 11:24:48 crc kubenswrapper[4824]: I1209 11:24:48.695745 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cw5vt" event={"ID":"7eab3325-7b88-47c2-b7cd-1a727a714b73","Type":"ContainerStarted","Data":"8a9edc7a12b4d4811920527ade45b7ba6e9a8a58fb2ea443f30de7b907e61edb"} Dec 09 11:24:48 crc kubenswrapper[4824]: I1209 11:24:48.703349 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 11:24:49 crc kubenswrapper[4824]: I1209 11:24:49.707881 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cw5vt" event={"ID":"7eab3325-7b88-47c2-b7cd-1a727a714b73","Type":"ContainerStarted","Data":"9d724269d41051d79c8506192225793e01fa0822c6775f7b581e073141684216"} Dec 09 11:24:50 crc kubenswrapper[4824]: I1209 11:24:50.723941 4824 generic.go:334] "Generic (PLEG): container finished" podID="7eab3325-7b88-47c2-b7cd-1a727a714b73" containerID="9d724269d41051d79c8506192225793e01fa0822c6775f7b581e073141684216" exitCode=0 Dec 09 11:24:50 crc kubenswrapper[4824]: I1209 11:24:50.724052 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cw5vt" event={"ID":"7eab3325-7b88-47c2-b7cd-1a727a714b73","Type":"ContainerDied","Data":"9d724269d41051d79c8506192225793e01fa0822c6775f7b581e073141684216"} Dec 09 11:24:51 crc kubenswrapper[4824]: I1209 11:24:51.753984 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cw5vt" event={"ID":"7eab3325-7b88-47c2-b7cd-1a727a714b73","Type":"ContainerStarted","Data":"f621bfd1732e1007ff675cb01af3978fb23af1ffdee81484f7025360a1116221"} Dec 09 11:24:51 crc kubenswrapper[4824]: I1209 11:24:51.785468 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cw5vt" podStartSLOduration=2.321479203 podStartE2EDuration="4.785440898s" podCreationTimestamp="2025-12-09 11:24:47 +0000 UTC" firstStartedPulling="2025-12-09 11:24:48.703099017 +0000 UTC m=+5845.037603684" lastFinishedPulling="2025-12-09 11:24:51.167060712 +0000 UTC m=+5847.501565379" observedRunningTime="2025-12-09 11:24:51.779955056 +0000 UTC m=+5848.114459723" watchObservedRunningTime="2025-12-09 11:24:51.785440898 +0000 UTC m=+5848.119945565" Dec 09 11:24:56 crc kubenswrapper[4824]: I1209 11:24:56.911624 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:24:56 crc kubenswrapper[4824]: E1209 11:24:56.914437 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:24:57 crc kubenswrapper[4824]: I1209 11:24:57.546372 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cw5vt" Dec 09 11:24:57 crc kubenswrapper[4824]: I1209 11:24:57.546433 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cw5vt" Dec 09 11:24:57 crc kubenswrapper[4824]: I1209 11:24:57.610713 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cw5vt" Dec 09 11:24:57 crc kubenswrapper[4824]: I1209 11:24:57.929181 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cw5vt" Dec 09 11:24:57 crc kubenswrapper[4824]: I1209 11:24:57.990211 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cw5vt"] Dec 09 11:24:59 crc kubenswrapper[4824]: I1209 11:24:59.873768 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cw5vt" podUID="7eab3325-7b88-47c2-b7cd-1a727a714b73" containerName="registry-server" containerID="cri-o://f621bfd1732e1007ff675cb01af3978fb23af1ffdee81484f7025360a1116221" gracePeriod=2 Dec 09 11:25:00 crc kubenswrapper[4824]: I1209 11:25:00.510676 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cw5vt" Dec 09 11:25:00 crc kubenswrapper[4824]: I1209 11:25:00.617852 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eab3325-7b88-47c2-b7cd-1a727a714b73-utilities\") pod \"7eab3325-7b88-47c2-b7cd-1a727a714b73\" (UID: \"7eab3325-7b88-47c2-b7cd-1a727a714b73\") " Dec 09 11:25:00 crc kubenswrapper[4824]: I1209 11:25:00.617974 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9fnw\" (UniqueName: \"kubernetes.io/projected/7eab3325-7b88-47c2-b7cd-1a727a714b73-kube-api-access-z9fnw\") pod \"7eab3325-7b88-47c2-b7cd-1a727a714b73\" (UID: \"7eab3325-7b88-47c2-b7cd-1a727a714b73\") " Dec 09 11:25:00 crc kubenswrapper[4824]: I1209 11:25:00.618093 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eab3325-7b88-47c2-b7cd-1a727a714b73-catalog-content\") pod \"7eab3325-7b88-47c2-b7cd-1a727a714b73\" (UID: \"7eab3325-7b88-47c2-b7cd-1a727a714b73\") " Dec 09 11:25:00 crc kubenswrapper[4824]: I1209 11:25:00.619042 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7eab3325-7b88-47c2-b7cd-1a727a714b73-utilities" (OuterVolumeSpecName: "utilities") pod "7eab3325-7b88-47c2-b7cd-1a727a714b73" (UID: "7eab3325-7b88-47c2-b7cd-1a727a714b73"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:25:00 crc kubenswrapper[4824]: I1209 11:25:00.624618 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7eab3325-7b88-47c2-b7cd-1a727a714b73-kube-api-access-z9fnw" (OuterVolumeSpecName: "kube-api-access-z9fnw") pod "7eab3325-7b88-47c2-b7cd-1a727a714b73" (UID: "7eab3325-7b88-47c2-b7cd-1a727a714b73"). InnerVolumeSpecName "kube-api-access-z9fnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 11:25:00 crc kubenswrapper[4824]: I1209 11:25:00.638737 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7eab3325-7b88-47c2-b7cd-1a727a714b73-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7eab3325-7b88-47c2-b7cd-1a727a714b73" (UID: "7eab3325-7b88-47c2-b7cd-1a727a714b73"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 11:25:00 crc kubenswrapper[4824]: I1209 11:25:00.721244 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eab3325-7b88-47c2-b7cd-1a727a714b73-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 11:25:00 crc kubenswrapper[4824]: I1209 11:25:00.721296 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eab3325-7b88-47c2-b7cd-1a727a714b73-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 11:25:00 crc kubenswrapper[4824]: I1209 11:25:00.721308 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9fnw\" (UniqueName: \"kubernetes.io/projected/7eab3325-7b88-47c2-b7cd-1a727a714b73-kube-api-access-z9fnw\") on node \"crc\" DevicePath \"\"" Dec 09 11:25:00 crc kubenswrapper[4824]: I1209 11:25:00.889371 4824 generic.go:334] "Generic (PLEG): container finished" podID="7eab3325-7b88-47c2-b7cd-1a727a714b73" containerID="f621bfd1732e1007ff675cb01af3978fb23af1ffdee81484f7025360a1116221" exitCode=0 Dec 09 11:25:00 crc kubenswrapper[4824]: I1209 11:25:00.889421 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cw5vt" event={"ID":"7eab3325-7b88-47c2-b7cd-1a727a714b73","Type":"ContainerDied","Data":"f621bfd1732e1007ff675cb01af3978fb23af1ffdee81484f7025360a1116221"} Dec 09 11:25:00 crc kubenswrapper[4824]: I1209 11:25:00.889450 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cw5vt" Dec 09 11:25:00 crc kubenswrapper[4824]: I1209 11:25:00.889470 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cw5vt" event={"ID":"7eab3325-7b88-47c2-b7cd-1a727a714b73","Type":"ContainerDied","Data":"8a9edc7a12b4d4811920527ade45b7ba6e9a8a58fb2ea443f30de7b907e61edb"} Dec 09 11:25:00 crc kubenswrapper[4824]: I1209 11:25:00.889496 4824 scope.go:117] "RemoveContainer" containerID="f621bfd1732e1007ff675cb01af3978fb23af1ffdee81484f7025360a1116221" Dec 09 11:25:00 crc kubenswrapper[4824]: I1209 11:25:00.918859 4824 scope.go:117] "RemoveContainer" containerID="9d724269d41051d79c8506192225793e01fa0822c6775f7b581e073141684216" Dec 09 11:25:00 crc kubenswrapper[4824]: I1209 11:25:00.932850 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cw5vt"] Dec 09 11:25:00 crc kubenswrapper[4824]: I1209 11:25:00.941892 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cw5vt"] Dec 09 11:25:00 crc kubenswrapper[4824]: I1209 11:25:00.950232 4824 scope.go:117] "RemoveContainer" containerID="493cc0a6fc4caf872a8d56dfd435fe2b27a8c288e308a6bac0a519b9b9674eab" Dec 09 11:25:01 crc kubenswrapper[4824]: I1209 11:25:01.016197 4824 scope.go:117] "RemoveContainer" containerID="f621bfd1732e1007ff675cb01af3978fb23af1ffdee81484f7025360a1116221" Dec 09 11:25:01 crc kubenswrapper[4824]: E1209 11:25:01.016647 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f621bfd1732e1007ff675cb01af3978fb23af1ffdee81484f7025360a1116221\": container with ID starting with f621bfd1732e1007ff675cb01af3978fb23af1ffdee81484f7025360a1116221 not found: ID does not exist" containerID="f621bfd1732e1007ff675cb01af3978fb23af1ffdee81484f7025360a1116221" Dec 09 11:25:01 crc kubenswrapper[4824]: I1209 11:25:01.016680 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f621bfd1732e1007ff675cb01af3978fb23af1ffdee81484f7025360a1116221"} err="failed to get container status \"f621bfd1732e1007ff675cb01af3978fb23af1ffdee81484f7025360a1116221\": rpc error: code = NotFound desc = could not find container \"f621bfd1732e1007ff675cb01af3978fb23af1ffdee81484f7025360a1116221\": container with ID starting with f621bfd1732e1007ff675cb01af3978fb23af1ffdee81484f7025360a1116221 not found: ID does not exist" Dec 09 11:25:01 crc kubenswrapper[4824]: I1209 11:25:01.016702 4824 scope.go:117] "RemoveContainer" containerID="9d724269d41051d79c8506192225793e01fa0822c6775f7b581e073141684216" Dec 09 11:25:01 crc kubenswrapper[4824]: E1209 11:25:01.017124 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d724269d41051d79c8506192225793e01fa0822c6775f7b581e073141684216\": container with ID starting with 9d724269d41051d79c8506192225793e01fa0822c6775f7b581e073141684216 not found: ID does not exist" containerID="9d724269d41051d79c8506192225793e01fa0822c6775f7b581e073141684216" Dec 09 11:25:01 crc kubenswrapper[4824]: I1209 11:25:01.017152 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d724269d41051d79c8506192225793e01fa0822c6775f7b581e073141684216"} err="failed to get container status \"9d724269d41051d79c8506192225793e01fa0822c6775f7b581e073141684216\": rpc error: code = NotFound desc = could not find container \"9d724269d41051d79c8506192225793e01fa0822c6775f7b581e073141684216\": container with ID starting with 9d724269d41051d79c8506192225793e01fa0822c6775f7b581e073141684216 not found: ID does not exist" Dec 09 11:25:01 crc kubenswrapper[4824]: I1209 11:25:01.017167 4824 scope.go:117] "RemoveContainer" containerID="493cc0a6fc4caf872a8d56dfd435fe2b27a8c288e308a6bac0a519b9b9674eab" Dec 09 11:25:01 crc kubenswrapper[4824]: E1209 11:25:01.017424 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"493cc0a6fc4caf872a8d56dfd435fe2b27a8c288e308a6bac0a519b9b9674eab\": container with ID starting with 493cc0a6fc4caf872a8d56dfd435fe2b27a8c288e308a6bac0a519b9b9674eab not found: ID does not exist" containerID="493cc0a6fc4caf872a8d56dfd435fe2b27a8c288e308a6bac0a519b9b9674eab" Dec 09 11:25:01 crc kubenswrapper[4824]: I1209 11:25:01.017450 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"493cc0a6fc4caf872a8d56dfd435fe2b27a8c288e308a6bac0a519b9b9674eab"} err="failed to get container status \"493cc0a6fc4caf872a8d56dfd435fe2b27a8c288e308a6bac0a519b9b9674eab\": rpc error: code = NotFound desc = could not find container \"493cc0a6fc4caf872a8d56dfd435fe2b27a8c288e308a6bac0a519b9b9674eab\": container with ID starting with 493cc0a6fc4caf872a8d56dfd435fe2b27a8c288e308a6bac0a519b9b9674eab not found: ID does not exist" Dec 09 11:25:01 crc kubenswrapper[4824]: I1209 11:25:01.927668 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7eab3325-7b88-47c2-b7cd-1a727a714b73" path="/var/lib/kubelet/pods/7eab3325-7b88-47c2-b7cd-1a727a714b73/volumes" Dec 09 11:25:10 crc kubenswrapper[4824]: I1209 11:25:10.910928 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:25:10 crc kubenswrapper[4824]: E1209 11:25:10.913775 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:25:21 crc kubenswrapper[4824]: I1209 11:25:21.914448 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:25:21 crc kubenswrapper[4824]: E1209 11:25:21.916564 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:25:36 crc kubenswrapper[4824]: I1209 11:25:36.911268 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:25:36 crc kubenswrapper[4824]: E1209 11:25:36.913968 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:25:50 crc kubenswrapper[4824]: I1209 11:25:50.910879 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:25:50 crc kubenswrapper[4824]: E1209 11:25:50.914009 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:26:01 crc kubenswrapper[4824]: I1209 11:26:01.911543 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:26:01 crc kubenswrapper[4824]: E1209 11:26:01.914280 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:26:12 crc kubenswrapper[4824]: I1209 11:26:12.911124 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:26:12 crc kubenswrapper[4824]: E1209 11:26:12.914650 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:26:26 crc kubenswrapper[4824]: I1209 11:26:26.911987 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:26:26 crc kubenswrapper[4824]: E1209 11:26:26.914689 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:26:39 crc kubenswrapper[4824]: I1209 11:26:39.911113 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:26:39 crc kubenswrapper[4824]: E1209 11:26:39.916141 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:26:52 crc kubenswrapper[4824]: I1209 11:26:52.911260 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:26:52 crc kubenswrapper[4824]: E1209 11:26:52.914010 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:27:02 crc kubenswrapper[4824]: I1209 11:27:02.860596 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:27:02 crc kubenswrapper[4824]: I1209 11:27:02.861196 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:27:04 crc kubenswrapper[4824]: I1209 11:27:04.911374 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:27:04 crc kubenswrapper[4824]: E1209 11:27:04.914312 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:27:15 crc kubenswrapper[4824]: I1209 11:27:15.914292 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:27:15 crc kubenswrapper[4824]: E1209 11:27:15.919523 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:27:26 crc kubenswrapper[4824]: I1209 11:27:26.911448 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:27:27 crc kubenswrapper[4824]: E1209 11:27:27.001537 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 11:27:27 crc kubenswrapper[4824]: E1209 11:27:27.002261 4824 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 11:27:27 crc kubenswrapper[4824]: E1209 11:27:27.002575 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n9h55h68ch65bh655h64chbfh6ch5b4h575h6bh654h5bdh587h5h58bh5dch5b9hf4h548h579hf8h5f5h686h667h97h57hb5h65h656h588h5f8q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6p7fx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9339c10c-1cea-4ffc-8931-9b349a1e74cd): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 11:27:27 crc kubenswrapper[4824]: E1209 11:27:27.005100 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:27:32 crc kubenswrapper[4824]: I1209 11:27:32.861035 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:27:32 crc kubenswrapper[4824]: I1209 11:27:32.861637 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:27:40 crc kubenswrapper[4824]: I1209 11:27:40.910725 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:27:40 crc kubenswrapper[4824]: E1209 11:27:40.913726 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:27:52 crc kubenswrapper[4824]: I1209 11:27:52.910571 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:27:52 crc kubenswrapper[4824]: E1209 11:27:52.913003 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:28:02 crc kubenswrapper[4824]: I1209 11:28:02.861201 4824 patch_prober.go:28] interesting pod/machine-config-daemon-dth8x container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 11:28:02 crc kubenswrapper[4824]: I1209 11:28:02.861918 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 11:28:02 crc kubenswrapper[4824]: I1209 11:28:02.861979 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" Dec 09 11:28:02 crc kubenswrapper[4824]: I1209 11:28:02.863067 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"34c807c5b217ed0d4c39e0fefa50a143310147aedb732f65dad4bdb215fe26a8"} pod="openshift-machine-config-operator/machine-config-daemon-dth8x" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 11:28:02 crc kubenswrapper[4824]: I1209 11:28:02.863138 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerName="machine-config-daemon" containerID="cri-o://34c807c5b217ed0d4c39e0fefa50a143310147aedb732f65dad4bdb215fe26a8" gracePeriod=600 Dec 09 11:28:02 crc kubenswrapper[4824]: E1209 11:28:02.994425 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:28:03 crc kubenswrapper[4824]: I1209 11:28:03.259323 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee3c86bf-f246-42cb-adb6-900e3d735166" containerID="34c807c5b217ed0d4c39e0fefa50a143310147aedb732f65dad4bdb215fe26a8" exitCode=0 Dec 09 11:28:03 crc kubenswrapper[4824]: I1209 11:28:03.259451 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" event={"ID":"ee3c86bf-f246-42cb-adb6-900e3d735166","Type":"ContainerDied","Data":"34c807c5b217ed0d4c39e0fefa50a143310147aedb732f65dad4bdb215fe26a8"} Dec 09 11:28:03 crc kubenswrapper[4824]: I1209 11:28:03.259939 4824 scope.go:117] "RemoveContainer" containerID="49a74dd2910658c46255de301a11501ac6eaebf752b480f0958816296c35a0f8" Dec 09 11:28:03 crc kubenswrapper[4824]: I1209 11:28:03.262104 4824 scope.go:117] "RemoveContainer" containerID="34c807c5b217ed0d4c39e0fefa50a143310147aedb732f65dad4bdb215fe26a8" Dec 09 11:28:03 crc kubenswrapper[4824]: E1209 11:28:03.262523 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:28:03 crc kubenswrapper[4824]: I1209 11:28:03.915683 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:28:03 crc kubenswrapper[4824]: E1209 11:28:03.917515 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:28:16 crc kubenswrapper[4824]: I1209 11:28:16.912721 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:28:16 crc kubenswrapper[4824]: E1209 11:28:16.918136 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:28:18 crc kubenswrapper[4824]: I1209 11:28:18.911465 4824 scope.go:117] "RemoveContainer" containerID="34c807c5b217ed0d4c39e0fefa50a143310147aedb732f65dad4bdb215fe26a8" Dec 09 11:28:18 crc kubenswrapper[4824]: E1209 11:28:18.912536 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:28:27 crc kubenswrapper[4824]: I1209 11:28:27.911508 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:28:27 crc kubenswrapper[4824]: E1209 11:28:27.914139 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:28:29 crc kubenswrapper[4824]: I1209 11:28:29.916642 4824 scope.go:117] "RemoveContainer" containerID="34c807c5b217ed0d4c39e0fefa50a143310147aedb732f65dad4bdb215fe26a8" Dec 09 11:28:29 crc kubenswrapper[4824]: E1209 11:28:29.920592 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:28:41 crc kubenswrapper[4824]: I1209 11:28:41.911798 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:28:41 crc kubenswrapper[4824]: E1209 11:28:41.915134 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:28:44 crc kubenswrapper[4824]: I1209 11:28:44.910892 4824 scope.go:117] "RemoveContainer" containerID="34c807c5b217ed0d4c39e0fefa50a143310147aedb732f65dad4bdb215fe26a8" Dec 09 11:28:44 crc kubenswrapper[4824]: E1209 11:28:44.911809 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:28:56 crc kubenswrapper[4824]: I1209 11:28:56.912156 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:28:56 crc kubenswrapper[4824]: E1209 11:28:56.916616 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:28:59 crc kubenswrapper[4824]: I1209 11:28:59.911713 4824 scope.go:117] "RemoveContainer" containerID="34c807c5b217ed0d4c39e0fefa50a143310147aedb732f65dad4bdb215fe26a8" Dec 09 11:28:59 crc kubenswrapper[4824]: E1209 11:28:59.912502 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:29:07 crc kubenswrapper[4824]: I1209 11:29:07.911213 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:29:07 crc kubenswrapper[4824]: E1209 11:29:07.914373 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:29:11 crc kubenswrapper[4824]: I1209 11:29:11.910629 4824 scope.go:117] "RemoveContainer" containerID="34c807c5b217ed0d4c39e0fefa50a143310147aedb732f65dad4bdb215fe26a8" Dec 09 11:29:11 crc kubenswrapper[4824]: E1209 11:29:11.911460 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" Dec 09 11:29:20 crc kubenswrapper[4824]: I1209 11:29:20.911129 4824 scope.go:117] "RemoveContainer" containerID="c574a8120337ff7539648ed3fca314ad0cf5d33f7b50d39e4e14f5e740d66fbd" Dec 09 11:29:20 crc kubenswrapper[4824]: E1209 11:29:20.914027 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9339c10c-1cea-4ffc-8931-9b349a1e74cd" Dec 09 11:29:24 crc kubenswrapper[4824]: I1209 11:29:24.911847 4824 scope.go:117] "RemoveContainer" containerID="34c807c5b217ed0d4c39e0fefa50a143310147aedb732f65dad4bdb215fe26a8" Dec 09 11:29:24 crc kubenswrapper[4824]: E1209 11:29:24.913125 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dth8x_openshift-machine-config-operator(ee3c86bf-f246-42cb-adb6-900e3d735166)\"" pod="openshift-machine-config-operator/machine-config-daemon-dth8x" podUID="ee3c86bf-f246-42cb-adb6-900e3d735166" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515116004045024441 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015116004045017356 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015115767546016526 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015115767546015476 5ustar corecore